From 05cf0d5cee36b3366388965bd40e7bc92ad78195 Mon Sep 17 00:00:00 2001 From: Mario Rodriguez Date: Thu, 30 Sep 2021 19:53:05 +0200 Subject: [PATCH 1/3] Upgrade jsonnet-libs to 1.19 and update tk examples --- CHANGELOG.md | 1 + example/tk/jsonnetfile.json | 6 +- example/tk/jsonnetfile.lock.json | 16 +- example/tk/lib/grafana/main.libsonnet | 6 +- example/tk/lib/k.libsonnet | 2 +- example/tk/lib/minio/minio.libsonnet | 2 +- example/tk/tempo-microservices/main.jsonnet | 20 +- example/tk/tempo-single-binary/main.jsonnet | 20 +- example/tk/vendor/1.19 | 1 + .../ksonnet-util/k-compat.libsonnet | 4 +- .../ksonnet-util/legacy-custom.libsonnet | 2 +- .../ksonnet-util/legacy-noname.libsonnet | 2 +- .../ksonnet-util/legacy-subtypes.libsonnet | 2 +- .../jsonnet-libs/ksonnet-util/util.libsonnet | 17 +- .../grafana/jsonnet-libs/memcached/README.md | 30 +- .../memcached/memcached.libsonnet | 5 +- .../k8s-libsonnet/1.19/_custom/apps.libsonnet | 83 + .../1.19/_custom/batch.libsonnet | 27 + .../k8s-libsonnet/1.19/_custom/core.libsonnet | 214 + .../k8s-libsonnet/1.19/_custom/list.libsonnet | 27 + .../1.19/_custom/mapContainers.libsonnet | 74 + .../k8s-libsonnet/1.19/_custom/rbac.libsonnet | 43 + .../1.19/_custom/volumeMounts.libsonnet | 169 + .../_gen/admissionregistration/main.libsonnet | 6 + .../admissionregistration/v1/main.libsonnet | 11 + .../v1/mutatingWebhook.libsonnet | 66 + .../v1/mutatingWebhookConfiguration.libsonnet | 60 + .../v1/ruleWithOperations.libsonnet | 24 + .../v1/serviceReference.libsonnet | 14 + .../v1/validatingWebhook.libsonnet | 64 + .../validatingWebhookConfiguration.libsonnet | 60 + .../v1/webhookClientConfig.libsonnet | 21 + .../v1beta1/main.libsonnet | 11 + .../v1beta1/mutatingWebhook.libsonnet | 66 + .../mutatingWebhookConfiguration.libsonnet | 60 + .../v1beta1/ruleWithOperations.libsonnet | 24 + .../v1beta1/serviceReference.libsonnet | 14 + .../v1beta1/validatingWebhook.libsonnet | 64 + .../validatingWebhookConfiguration.libsonnet | 60 + .../v1beta1/webhookClientConfig.libsonnet | 21 + .../1.19/_gen/apiregistration/main.libsonnet | 6 + .../apiregistration/v1/apiService.libsonnet | 80 + .../v1/apiServiceCondition.libsonnet | 14 + .../v1/apiServiceSpec.libsonnet | 27 + .../v1/apiServiceStatus.libsonnet | 10 + .../_gen/apiregistration/v1/main.libsonnet | 9 + .../v1/serviceReference.libsonnet | 12 + .../v1beta1/apiService.libsonnet | 80 + .../v1beta1/apiServiceCondition.libsonnet | 14 + .../v1beta1/apiServiceSpec.libsonnet | 27 + .../v1beta1/apiServiceStatus.libsonnet | 10 + .../apiregistration/v1beta1/main.libsonnet | 9 + .../v1beta1/serviceReference.libsonnet | 12 + .../1.19/_gen/apps/main.libsonnet | 5 + .../_gen/apps/v1/controllerRevision.libsonnet | 62 + .../1.19/_gen/apps/v1/daemonSet.libsonnet | 323 + .../_gen/apps/v1/daemonSetCondition.libsonnet | 14 + .../1.19/_gen/apps/v1/daemonSetSpec.libsonnet | 270 + .../_gen/apps/v1/daemonSetStatus.libsonnet | 28 + .../apps/v1/daemonSetUpdateStrategy.libsonnet | 13 + .../1.19/_gen/apps/v1/deployment.libsonnet | 331 + .../apps/v1/deploymentCondition.libsonnet | 16 + .../_gen/apps/v1/deploymentSpec.libsonnet | 278 + .../_gen/apps/v1/deploymentStatus.libsonnet | 24 + .../_gen/apps/v1/deploymentStrategy.libsonnet | 15 + .../1.19/_gen/apps/v1/main.libsonnet | 27 + .../1.19/_gen/apps/v1/replicaSet.libsonnet | 313 + .../apps/v1/replicaSetCondition.libsonnet | 14 + .../_gen/apps/v1/replicaSetSpec.libsonnet | 260 + .../_gen/apps/v1/replicaSetStatus.libsonnet | 20 + .../apps/v1/rollingUpdateDaemonSet.libsonnet | 8 + .../apps/v1/rollingUpdateDeployment.libsonnet | 10 + ...rollingUpdateStatefulSetStrategy.libsonnet | 8 + .../1.19/_gen/apps/v1/statefulSet.libsonnet | 331 + .../apps/v1/statefulSetCondition.libsonnet | 14 + .../_gen/apps/v1/statefulSetSpec.libsonnet | 278 + .../_gen/apps/v1/statefulSetStatus.libsonnet | 26 + .../v1/statefulSetUpdateStrategy.libsonnet | 13 + .../1.19/_gen/authentication/main.libsonnet | 6 + .../v1/boundObjectReference.libsonnet | 12 + .../_gen/authentication/v1/main.libsonnet | 12 + .../authentication/v1/tokenRequest.libsonnet | 76 + .../v1/tokenRequestSpec.libsonnet | 23 + .../v1/tokenRequestStatus.libsonnet | 10 + .../authentication/v1/tokenReview.libsonnet | 65 + .../v1/tokenReviewSpec.libsonnet | 12 + .../v1/tokenReviewStatus.libsonnet | 29 + .../_gen/authentication/v1/userInfo.libsonnet | 18 + .../authentication/v1beta1/main.libsonnet | 8 + .../v1beta1/tokenReview.libsonnet | 65 + .../v1beta1/tokenReviewSpec.libsonnet | 12 + .../v1beta1/tokenReviewStatus.libsonnet | 29 + .../authentication/v1beta1/userInfo.libsonnet | 18 + .../1.19/_gen/authorization/main.libsonnet | 6 + .../v1/localSubjectAccessReview.libsonnet | 95 + .../1.19/_gen/authorization/v1/main.libsonnet | 17 + .../v1/nonResourceAttributes.libsonnet | 10 + .../v1/nonResourceRule.libsonnet | 14 + .../v1/resourceAttributes.libsonnet | 20 + .../authorization/v1/resourceRule.libsonnet | 22 + .../v1/selfSubjectAccessReview.libsonnet | 83 + .../v1/selfSubjectAccessReviewSpec.libsonnet | 30 + .../v1/selfSubjectRulesReview.libsonnet | 61 + .../v1/selfSubjectRulesReviewSpec.libsonnet | 8 + .../v1/subjectAccessReview.libsonnet | 95 + .../v1/subjectAccessReviewSpec.libsonnet | 42 + .../v1/subjectAccessReviewStatus.libsonnet | 14 + .../v1/subjectRulesReviewStatus.libsonnet | 18 + .../localSubjectAccessReview.libsonnet | 95 + .../_gen/authorization/v1beta1/main.libsonnet | 17 + .../v1beta1/nonResourceAttributes.libsonnet | 10 + .../v1beta1/nonResourceRule.libsonnet | 14 + .../v1beta1/resourceAttributes.libsonnet | 20 + .../v1beta1/resourceRule.libsonnet | 22 + .../v1beta1/selfSubjectAccessReview.libsonnet | 83 + .../selfSubjectAccessReviewSpec.libsonnet | 30 + .../v1beta1/selfSubjectRulesReview.libsonnet | 61 + .../selfSubjectRulesReviewSpec.libsonnet | 8 + .../v1beta1/subjectAccessReview.libsonnet | 95 + .../v1beta1/subjectAccessReviewSpec.libsonnet | 42 + .../subjectAccessReviewStatus.libsonnet | 14 + .../subjectRulesReviewStatus.libsonnet | 18 + .../1.19/_gen/autoscaling/main.libsonnet | 7 + .../v1/crossVersionObjectReference.libsonnet | 10 + .../v1/horizontalPodAutoscaler.libsonnet | 74 + .../v1/horizontalPodAutoscalerSpec.libsonnet | 21 + .../horizontalPodAutoscalerStatus.libsonnet | 16 + .../1.19/_gen/autoscaling/v1/main.libsonnet | 11 + .../1.19/_gen/autoscaling/v1/scale.libsonnet | 61 + .../_gen/autoscaling/v1/scaleSpec.libsonnet | 8 + .../_gen/autoscaling/v1/scaleStatus.libsonnet | 10 + .../crossVersionObjectReference.libsonnet | 10 + .../v2beta1/externalMetricSource.libsonnet | 23 + .../v2beta1/externalMetricStatus.libsonnet | 23 + .../v2beta1/horizontalPodAutoscaler.libsonnet | 76 + ...horizontalPodAutoscalerCondition.libsonnet | 14 + .../horizontalPodAutoscalerSpec.libsonnet | 23 + .../horizontalPodAutoscalerStatus.libsonnet | 22 + .../_gen/autoscaling/v2beta1/main.libsonnet | 19 + .../autoscaling/v2beta1/metricSpec.libsonnet | 84 + .../v2beta1/metricStatus.libsonnet | 84 + .../v2beta1/objectMetricSource.libsonnet | 32 + .../v2beta1/objectMetricStatus.libsonnet | 32 + .../v2beta1/podsMetricSource.libsonnet | 21 + .../v2beta1/podsMetricStatus.libsonnet | 21 + .../v2beta1/resourceMetricSource.libsonnet | 12 + .../v2beta1/resourceMetricStatus.libsonnet | 12 + .../crossVersionObjectReference.libsonnet | 10 + .../v2beta2/externalMetricSource.libsonnet | 33 + .../v2beta2/externalMetricStatus.libsonnet | 31 + .../v2beta2/horizontalPodAutoscaler.libsonnet | 101 + .../horizontalPodAutoscalerBehavior.libsonnet | 28 + ...horizontalPodAutoscalerCondition.libsonnet | 14 + .../horizontalPodAutoscalerSpec.libsonnet | 48 + .../horizontalPodAutoscalerStatus.libsonnet | 22 + .../v2beta2/hpaScalingPolicy.libsonnet | 12 + .../v2beta2/hpaScalingRules.libsonnet | 14 + .../_gen/autoscaling/v2beta2/main.libsonnet | 25 + .../v2beta2/metricIdentifier.libsonnet | 19 + .../autoscaling/v2beta2/metricSpec.libsonnet | 123 + .../v2beta2/metricStatus.libsonnet | 115 + .../v2beta2/metricTarget.libsonnet | 14 + .../v2beta2/metricValueStatus.libsonnet | 12 + .../v2beta2/objectMetricSource.libsonnet | 42 + .../v2beta2/objectMetricStatus.libsonnet | 40 + .../v2beta2/podsMetricSource.libsonnet | 33 + .../v2beta2/podsMetricStatus.libsonnet | 31 + .../v2beta2/resourceMetricSource.libsonnet | 19 + .../v2beta2/resourceMetricStatus.libsonnet | 17 + .../1.19/_gen/batch/main.libsonnet | 7 + .../1.19/_gen/batch/v1/job.libsonnet | 321 + .../1.19/_gen/batch/v1/jobCondition.libsonnet | 16 + .../1.19/_gen/batch/v1/jobSpec.libsonnet | 268 + .../1.19/_gen/batch/v1/jobStatus.libsonnet | 20 + .../1.19/_gen/batch/v1/main.libsonnet | 8 + .../1.19/_gen/batch/v1beta1/cronJob.libsonnet | 384 + .../_gen/batch/v1beta1/cronJobSpec.libsonnet | 331 + .../batch/v1beta1/cronJobStatus.libsonnet | 12 + .../batch/v1beta1/jobTemplateSpec.libsonnet | 316 + .../1.19/_gen/batch/v1beta1/main.libsonnet | 8 + .../_gen/batch/v2alpha1/cronJob.libsonnet | 384 + .../_gen/batch/v2alpha1/cronJobSpec.libsonnet | 331 + .../batch/v2alpha1/cronJobStatus.libsonnet | 12 + .../batch/v2alpha1/jobTemplateSpec.libsonnet | 316 + .../1.19/_gen/batch/v2alpha1/main.libsonnet | 8 + .../1.19/_gen/certificates/main.libsonnet | 6 + .../v1/certificateSigningRequest.libsonnet | 79 + ...rtificateSigningRequestCondition.libsonnet | 16 + .../certificateSigningRequestSpec.libsonnet | 26 + .../certificateSigningRequestStatus.libsonnet | 12 + .../1.19/_gen/certificates/v1/main.libsonnet | 8 + .../certificateSigningRequest.libsonnet | 79 + ...rtificateSigningRequestCondition.libsonnet | 16 + .../certificateSigningRequestSpec.libsonnet | 26 + .../certificateSigningRequestStatus.libsonnet | 12 + .../_gen/certificates/v1beta1/main.libsonnet | 8 + .../1.19/_gen/coordination/main.libsonnet | 6 + .../1.19/_gen/coordination/v1/lease.libsonnet | 69 + .../_gen/coordination/v1/leaseSpec.libsonnet | 16 + .../1.19/_gen/coordination/v1/main.libsonnet | 6 + .../_gen/coordination/v1beta1/lease.libsonnet | 69 + .../coordination/v1beta1/leaseSpec.libsonnet | 16 + .../_gen/coordination/v1beta1/main.libsonnet | 6 + .../1.19/_gen/core/main.libsonnet | 5 + .../1.19/_gen/core/v1/affinity.libsonnet | 42 + .../_gen/core/v1/attachedVolume.libsonnet | 10 + ...awsElasticBlockStoreVolumeSource.libsonnet | 14 + .../core/v1/azureDiskVolumeSource.libsonnet | 18 + .../azureFilePersistentVolumeSource.libsonnet | 14 + .../core/v1/azureFileVolumeSource.libsonnet | 12 + .../1.19/_gen/core/v1/binding.libsonnet | 73 + .../1.19/_gen/core/v1/capabilities.libsonnet | 14 + .../v1/cephFSPersistentVolumeSource.libsonnet | 25 + .../_gen/core/v1/cephFSVolumeSource.libsonnet | 23 + .../v1/cinderPersistentVolumeSource.libsonnet | 19 + .../_gen/core/v1/cinderVolumeSource.libsonnet | 17 + .../_gen/core/v1/clientIPConfig.libsonnet | 8 + .../_gen/core/v1/componentCondition.libsonnet | 12 + .../_gen/core/v1/componentStatus.libsonnet | 60 + .../1.19/_gen/core/v1/configMap.libsonnet | 66 + .../_gen/core/v1/configMapEnvSource.libsonnet | 10 + .../core/v1/configMapKeySelector.libsonnet | 12 + .../v1/configMapNodeConfigSource.libsonnet | 16 + .../core/v1/configMapProjection.libsonnet | 14 + .../core/v1/configMapVolumeSource.libsonnet | 16 + .../1.19/_gen/core/v1/container.libsonnet | 311 + .../_gen/core/v1/containerImage.libsonnet | 12 + .../1.19/_gen/core/v1/containerPort.libsonnet | 16 + .../_gen/core/v1/containerState.libsonnet | 35 + .../core/v1/containerStateRunning.libsonnet | 8 + .../v1/containerStateTerminated.libsonnet | 20 + .../core/v1/containerStateWaiting.libsonnet | 10 + .../_gen/core/v1/containerStatus.libsonnet | 84 + .../v1/csiPersistentVolumeSource.libsonnet | 46 + .../_gen/core/v1/csiVolumeSource.libsonnet | 21 + .../_gen/core/v1/daemonEndpoint.libsonnet | 8 + .../core/v1/downwardAPIProjection.libsonnet | 10 + .../core/v1/downwardAPIVolumeFile.libsonnet | 26 + .../core/v1/downwardAPIVolumeSource.libsonnet | 12 + .../core/v1/emptyDirVolumeSource.libsonnet | 10 + .../_gen/core/v1/endpointAddress.libsonnet | 29 + .../1.19/_gen/core/v1/endpointPort.libsonnet | 14 + .../_gen/core/v1/endpointSubset.libsonnet | 18 + .../1.19/_gen/core/v1/endpoints.libsonnet | 60 + .../1.19/_gen/core/v1/envFromSource.libsonnet | 22 + .../1.19/_gen/core/v1/envVar.libsonnet | 47 + .../1.19/_gen/core/v1/envVarSource.libsonnet | 40 + .../_gen/core/v1/ephemeralContainer.libsonnet | 313 + .../core/v1/ephemeralVolumeSource.libsonnet | 100 + .../1.19/_gen/core/v1/event.libsonnet | 124 + .../1.19/_gen/core/v1/eventSeries.libsonnet | 10 + .../1.19/_gen/core/v1/eventSource.libsonnet | 10 + .../1.19/_gen/core/v1/execAction.libsonnet | 10 + .../_gen/core/v1/fcVolumeSource.libsonnet | 20 + .../v1/flexPersistentVolumeSource.libsonnet | 23 + .../_gen/core/v1/flexVolumeSource.libsonnet | 21 + .../core/v1/flockerVolumeSource.libsonnet | 10 + .../gcePersistentDiskVolumeSource.libsonnet | 14 + .../core/v1/gitRepoVolumeSource.libsonnet | 12 + .../glusterfsPersistentVolumeSource.libsonnet | 14 + .../core/v1/glusterfsVolumeSource.libsonnet | 12 + .../1.19/_gen/core/v1/handler.libsonnet | 35 + .../1.19/_gen/core/v1/hostAlias.libsonnet | 12 + .../core/v1/hostPathVolumeSource.libsonnet | 10 + .../1.19/_gen/core/v1/httpGetAction.libsonnet | 18 + .../1.19/_gen/core/v1/httpHeader.libsonnet | 10 + .../v1/iscsiPersistentVolumeSource.libsonnet | 35 + .../_gen/core/v1/iscsiVolumeSource.libsonnet | 33 + .../1.19/_gen/core/v1/keyToPath.libsonnet | 12 + .../1.19/_gen/core/v1/lifecycle.libsonnet | 70 + .../1.19/_gen/core/v1/limitRange.libsonnet | 63 + .../_gen/core/v1/limitRangeItem.libsonnet | 28 + .../_gen/core/v1/limitRangeSpec.libsonnet | 10 + .../core/v1/loadBalancerIngress.libsonnet | 10 + .../_gen/core/v1/loadBalancerStatus.libsonnet | 10 + .../core/v1/localObjectReference.libsonnet | 8 + .../_gen/core/v1/localVolumeSource.libsonnet | 10 + .../1.19/_gen/core/v1/main.libsonnet | 175 + .../1.19/_gen/core/v1/namespace.libsonnet | 63 + .../_gen/core/v1/namespaceCondition.libsonnet | 14 + .../1.19/_gen/core/v1/namespaceSpec.libsonnet | 10 + .../_gen/core/v1/namespaceStatus.libsonnet | 12 + .../_gen/core/v1/nfsVolumeSource.libsonnet | 12 + .../1.19/_gen/core/v1/node.libsonnet | 91 + .../1.19/_gen/core/v1/nodeAddress.libsonnet | 10 + .../1.19/_gen/core/v1/nodeAffinity.libsonnet | 17 + .../1.19/_gen/core/v1/nodeCondition.libsonnet | 16 + .../_gen/core/v1/nodeConfigSource.libsonnet | 19 + .../_gen/core/v1/nodeConfigStatus.libsonnet | 56 + .../core/v1/nodeDaemonEndpoints.libsonnet | 11 + .../1.19/_gen/core/v1/nodeSelector.libsonnet | 10 + .../core/v1/nodeSelectorRequirement.libsonnet | 14 + .../_gen/core/v1/nodeSelectorTerm.libsonnet | 14 + .../1.19/_gen/core/v1/nodeSpec.libsonnet | 38 + .../1.19/_gen/core/v1/nodeStatus.libsonnet | 120 + .../_gen/core/v1/nodeSystemInfo.libsonnet | 26 + .../core/v1/objectFieldSelector.libsonnet | 8 + .../_gen/core/v1/objectReference.libsonnet | 18 + .../_gen/core/v1/persistentVolume.libsonnet | 467 + .../core/v1/persistentVolumeClaim.libsonnet | 100 + .../persistentVolumeClaimCondition.libsonnet | 16 + .../v1/persistentVolumeClaimSpec.libsonnet | 47 + .../v1/persistentVolumeClaimStatus.libsonnet | 20 + .../persistentVolumeClaimTemplate.libsonnet | 95 + ...ersistentVolumeClaimVolumeSource.libsonnet | 10 + .../core/v1/persistentVolumeSpec.libsonnet | 414 + .../core/v1/persistentVolumeStatus.libsonnet | 12 + ...photonPersistentDiskVolumeSource.libsonnet | 10 + .../1.19/_gen/core/v1/pod.libsonnet | 247 + .../1.19/_gen/core/v1/podAffinity.libsonnet | 14 + .../_gen/core/v1/podAffinityTerm.libsonnet | 23 + .../_gen/core/v1/podAntiAffinity.libsonnet | 14 + .../1.19/_gen/core/v1/podCondition.libsonnet | 16 + .../1.19/_gen/core/v1/podDNSConfig.libsonnet | 18 + .../_gen/core/v1/podDNSConfigOption.libsonnet | 10 + .../1.19/_gen/core/v1/podIP.libsonnet | 8 + .../_gen/core/v1/podReadinessGate.libsonnet | 8 + .../_gen/core/v1/podSecurityContext.libsonnet | 51 + .../1.19/_gen/core/v1/podSpec.libsonnet | 194 + .../1.19/_gen/core/v1/podStatus.libsonnet | 42 + .../1.19/_gen/core/v1/podTemplate.libsonnet | 295 + .../_gen/core/v1/podTemplateSpec.libsonnet | 242 + .../core/v1/portworxVolumeSource.libsonnet | 12 + .../core/v1/preferredSchedulingTerm.libsonnet | 19 + .../1.19/_gen/core/v1/probe.libsonnet | 45 + .../core/v1/projectedVolumeSource.libsonnet | 12 + .../core/v1/quobyteVolumeSource.libsonnet | 18 + .../v1/rbdPersistentVolumeSource.libsonnet | 29 + .../_gen/core/v1/rbdVolumeSource.libsonnet | 27 + .../core/v1/replicationController.libsonnet | 306 + .../replicationControllerCondition.libsonnet | 14 + .../v1/replicationControllerSpec.libsonnet | 253 + .../v1/replicationControllerStatus.libsonnet | 20 + .../core/v1/resourceFieldSelector.libsonnet | 12 + .../1.19/_gen/core/v1/resourceQuota.libsonnet | 74 + .../_gen/core/v1/resourceQuotaSpec.libsonnet | 21 + .../core/v1/resourceQuotaStatus.libsonnet | 14 + .../core/v1/resourceRequirements.libsonnet | 14 + .../scaleIOPersistentVolumeSource.libsonnet | 31 + .../core/v1/scaleIOVolumeSource.libsonnet | 29 + .../1.19/_gen/core/v1/scopeSelector.libsonnet | 10 + ...copedResourceSelectorRequirement.libsonnet | 14 + .../_gen/core/v1/seLinuxOptions.libsonnet | 14 + .../_gen/core/v1/seccompProfile.libsonnet | 10 + .../1.19/_gen/core/v1/secret.libsonnet | 68 + .../_gen/core/v1/secretEnvSource.libsonnet | 10 + .../_gen/core/v1/secretKeySelector.libsonnet | 12 + .../_gen/core/v1/secretProjection.libsonnet | 14 + .../_gen/core/v1/secretReference.libsonnet | 10 + .../_gen/core/v1/secretVolumeSource.libsonnet | 16 + .../_gen/core/v1/securityContext.libsonnet | 58 + .../1.19/_gen/core/v1/service.libsonnet | 105 + .../_gen/core/v1/serviceAccount.libsonnet | 66 + .../serviceAccountTokenProjection.libsonnet | 12 + .../1.19/_gen/core/v1/servicePort.libsonnet | 18 + .../1.19/_gen/core/v1/serviceSpec.libsonnet | 52 + .../1.19/_gen/core/v1/serviceStatus.libsonnet | 13 + .../core/v1/sessionAffinityConfig.libsonnet | 11 + .../storageOSPersistentVolumeSource.libsonnet | 31 + .../core/v1/storageOSVolumeSource.libsonnet | 19 + .../1.19/_gen/core/v1/sysctl.libsonnet | 10 + .../1.19/_gen/core/v1/taint.libsonnet | 14 + .../_gen/core/v1/tcpSocketAction.libsonnet | 10 + .../1.19/_gen/core/v1/toleration.libsonnet | 16 + ...topologySelectorLabelRequirement.libsonnet | 12 + .../core/v1/topologySelectorTerm.libsonnet | 10 + .../v1/topologySpreadConstraint.libsonnet | 23 + .../v1/typedLocalObjectReference.libsonnet | 12 + .../1.19/_gen/core/v1/volume.libsonnet | 475 + .../1.19/_gen/core/v1/volumeDevice.libsonnet | 10 + .../1.19/_gen/core/v1/volumeMount.libsonnet | 18 + .../_gen/core/v1/volumeNodeAffinity.libsonnet | 13 + .../_gen/core/v1/volumeProjection.libsonnet | 44 + .../vsphereVirtualDiskVolumeSource.libsonnet | 14 + .../core/v1/weightedPodAffinityTerm.libsonnet | 28 + .../windowsSecurityContextOptions.libsonnet | 12 + .../1.19/_gen/discovery/main.libsonnet | 5 + .../_gen/discovery/v1beta1/endpoint.libsonnet | 38 + .../v1beta1/endpointConditions.libsonnet | 8 + .../discovery/v1beta1/endpointPort.libsonnet | 14 + .../discovery/v1beta1/endpointSlice.libsonnet | 66 + .../_gen/discovery/v1beta1/main.libsonnet | 8 + .../1.19/_gen/events/main.libsonnet | 6 + .../1.19/_gen/events/v1/event.libsonnet | 124 + .../1.19/_gen/events/v1/eventSeries.libsonnet | 10 + .../1.19/_gen/events/v1/main.libsonnet | 6 + .../1.19/_gen/events/v1beta1/event.libsonnet | 124 + .../_gen/events/v1beta1/eventSeries.libsonnet | 10 + .../1.19/_gen/events/v1beta1/main.libsonnet | 6 + .../1.19/_gen/extensions/main.libsonnet | 5 + .../v1beta1/httpIngressPath.libsonnet | 26 + .../v1beta1/httpIngressRuleValue.libsonnet | 10 + .../_gen/extensions/v1beta1/ingress.libsonnet | 85 + .../v1beta1/ingressBackend.libsonnet | 19 + .../extensions/v1beta1/ingressRule.libsonnet | 15 + .../extensions/v1beta1/ingressSpec.libsonnet | 32 + .../v1beta1/ingressStatus.libsonnet | 13 + .../extensions/v1beta1/ingressTLS.libsonnet | 12 + .../_gen/extensions/v1beta1/main.libsonnet | 12 + .../1.19/_gen/flowcontrol/main.libsonnet | 5 + .../flowDistinguisherMethod.libsonnet | 8 + .../flowcontrol/v1alpha1/flowSchema.libsonnet | 75 + .../v1alpha1/flowSchemaCondition.libsonnet | 14 + .../v1alpha1/flowSchemaSpec.libsonnet | 22 + .../v1alpha1/flowSchemaStatus.libsonnet | 10 + .../v1alpha1/groupSubject.libsonnet | 8 + .../v1alpha1/limitResponse.libsonnet | 17 + ...imitedPriorityLevelConfiguration.libsonnet | 22 + .../_gen/flowcontrol/v1alpha1/main.libsonnet | 24 + .../v1alpha1/nonResourcePolicyRule.libsonnet | 14 + .../policyRulesWithSubjects.libsonnet | 18 + .../priorityLevelConfiguration.libsonnet | 80 + ...orityLevelConfigurationCondition.libsonnet | 14 + ...orityLevelConfigurationReference.libsonnet | 8 + .../priorityLevelConfigurationSpec.libsonnet | 27 + ...priorityLevelConfigurationStatus.libsonnet | 10 + .../v1alpha1/queuingConfiguration.libsonnet | 12 + .../v1alpha1/resourcePolicyRule.libsonnet | 24 + .../v1alpha1/serviceAccountSubject.libsonnet | 10 + .../flowcontrol/v1alpha1/subject.libsonnet | 25 + .../v1alpha1/userSubject.libsonnet | 8 + .../1.19/_gen/networking/main.libsonnet | 6 + .../networking/v1/httpIngressPath.libsonnet | 34 + .../v1/httpIngressRuleValue.libsonnet | 10 + .../1.19/_gen/networking/v1/ingress.libsonnet | 93 + .../networking/v1/ingressBackend.libsonnet | 27 + .../_gen/networking/v1/ingressClass.libsonnet | 70 + .../networking/v1/ingressClassSpec.libsonnet | 17 + .../_gen/networking/v1/ingressRule.libsonnet | 15 + .../v1/ingressServiceBackend.libsonnet | 15 + .../_gen/networking/v1/ingressSpec.libsonnet | 40 + .../networking/v1/ingressStatus.libsonnet | 13 + .../_gen/networking/v1/ingressTLS.libsonnet | 12 + .../1.19/_gen/networking/v1/ipBlock.libsonnet | 12 + .../1.19/_gen/networking/v1/main.libsonnet | 23 + .../networking/v1/networkPolicy.libsonnet | 82 + .../v1/networkPolicyEgressRule.libsonnet | 14 + .../v1/networkPolicyIngressRule.libsonnet | 14 + .../networking/v1/networkPolicyPeer.libsonnet | 37 + .../networking/v1/networkPolicyPort.libsonnet | 10 + .../networking/v1/networkPolicySpec.libsonnet | 29 + .../v1/serviceBackendPort.libsonnet | 10 + .../v1beta1/httpIngressPath.libsonnet | 26 + .../v1beta1/httpIngressRuleValue.libsonnet | 10 + .../_gen/networking/v1beta1/ingress.libsonnet | 85 + .../v1beta1/ingressBackend.libsonnet | 19 + .../networking/v1beta1/ingressClass.libsonnet | 70 + .../v1beta1/ingressClassSpec.libsonnet | 17 + .../networking/v1beta1/ingressRule.libsonnet | 15 + .../networking/v1beta1/ingressSpec.libsonnet | 32 + .../v1beta1/ingressStatus.libsonnet | 13 + .../networking/v1beta1/ingressTLS.libsonnet | 12 + .../_gen/networking/v1beta1/main.libsonnet | 14 + .../1.19/_gen/node/main.libsonnet | 6 + .../1.19/_gen/node/v1alpha1/main.libsonnet | 8 + .../_gen/node/v1alpha1/overhead.libsonnet | 10 + .../_gen/node/v1alpha1/runtimeClass.libsonnet | 79 + .../node/v1alpha1/runtimeClassSpec.libsonnet | 26 + .../_gen/node/v1alpha1/scheduling.libsonnet | 14 + .../1.19/_gen/node/v1beta1/main.libsonnet | 7 + .../1.19/_gen/node/v1beta1/overhead.libsonnet | 10 + .../_gen/node/v1beta1/runtimeClass.libsonnet | 76 + .../_gen/node/v1beta1/scheduling.libsonnet | 14 + .../1.19/_gen/policy/main.libsonnet | 5 + .../policy/v1beta1/allowedCSIDriver.libsonnet | 8 + .../v1beta1/allowedFlexVolume.libsonnet | 8 + .../policy/v1beta1/allowedHostPath.libsonnet | 10 + .../_gen/policy/v1beta1/eviction.libsonnet | 80 + .../v1beta1/fsGroupStrategyOptions.libsonnet | 12 + .../policy/v1beta1/hostPortRange.libsonnet | 10 + .../_gen/policy/v1beta1/idRange.libsonnet | 10 + .../1.19/_gen/policy/v1beta1/main.libsonnet | 21 + .../v1beta1/podDisruptionBudget.libsonnet | 74 + .../v1beta1/podDisruptionBudgetSpec.libsonnet | 21 + .../podDisruptionBudgetStatus.libsonnet | 20 + .../v1beta1/podSecurityPolicy.libsonnet | 178 + .../v1beta1/podSecurityPolicySpec.libsonnet | 125 + .../runAsGroupStrategyOptions.libsonnet | 12 + .../runAsUserStrategyOptions.libsonnet | 12 + .../runtimeClassStrategyOptions.libsonnet | 12 + .../v1beta1/seLinuxStrategyOptions.libsonnet | 19 + ...upplementalGroupsStrategyOptions.libsonnet | 12 + .../1.19/_gen/rbac/main.libsonnet | 7 + .../_gen/rbac/v1/aggregationRule.libsonnet | 10 + .../1.19/_gen/rbac/v1/clusterRole.libsonnet | 67 + .../_gen/rbac/v1/clusterRoleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1/main.libsonnet | 12 + .../1.19/_gen/rbac/v1/policyRule.libsonnet | 26 + .../1.19/_gen/rbac/v1/role.libsonnet | 60 + .../1.19/_gen/rbac/v1/roleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1/roleRef.libsonnet | 12 + .../1.19/_gen/rbac/v1/subject.libsonnet | 14 + .../rbac/v1alpha1/aggregationRule.libsonnet | 10 + .../_gen/rbac/v1alpha1/clusterRole.libsonnet | 67 + .../v1alpha1/clusterRoleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1alpha1/main.libsonnet | 12 + .../_gen/rbac/v1alpha1/policyRule.libsonnet | 26 + .../1.19/_gen/rbac/v1alpha1/role.libsonnet | 60 + .../_gen/rbac/v1alpha1/roleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1alpha1/roleRef.libsonnet | 12 + .../1.19/_gen/rbac/v1alpha1/subject.libsonnet | 12 + .../rbac/v1beta1/aggregationRule.libsonnet | 10 + .../_gen/rbac/v1beta1/clusterRole.libsonnet | 67 + .../rbac/v1beta1/clusterRoleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1beta1/main.libsonnet | 12 + .../_gen/rbac/v1beta1/policyRule.libsonnet | 26 + .../1.19/_gen/rbac/v1beta1/role.libsonnet | 60 + .../_gen/rbac/v1beta1/roleBinding.libsonnet | 69 + .../1.19/_gen/rbac/v1beta1/roleRef.libsonnet | 12 + .../1.19/_gen/rbac/v1beta1/subject.libsonnet | 14 + .../1.19/_gen/scheduling/main.libsonnet | 7 + .../1.19/_gen/scheduling/v1/main.libsonnet | 5 + .../scheduling/v1/priorityClass.libsonnet | 64 + .../_gen/scheduling/v1alpha1/main.libsonnet | 5 + .../v1alpha1/priorityClass.libsonnet | 64 + .../_gen/scheduling/v1beta1/main.libsonnet | 5 + .../v1beta1/priorityClass.libsonnet | 64 + .../1.19/_gen/settings/main.libsonnet | 5 + .../_gen/settings/v1alpha1/main.libsonnet | 6 + .../settings/v1alpha1/podPreset.libsonnet | 86 + .../settings/v1alpha1/podPresetSpec.libsonnet | 33 + .../1.19/_gen/storage/main.libsonnet | 7 + .../1.19/_gen/storage/v1/csiDriver.libsonnet | 71 + .../_gen/storage/v1/csiDriverSpec.libsonnet | 18 + .../1.19/_gen/storage/v1/csiNode.libsonnet | 63 + .../_gen/storage/v1/csiNodeDriver.libsonnet | 19 + .../_gen/storage/v1/csiNodeSpec.libsonnet | 10 + .../1.19/_gen/storage/v1/main.libsonnet | 16 + .../_gen/storage/v1/storageClass.libsonnet | 76 + .../storage/v1/volumeAttachment.libsonnet | 479 + .../v1/volumeAttachmentSource.libsonnet | 419 + .../storage/v1/volumeAttachmentSpec.libsonnet | 426 + .../v1/volumeAttachmentStatus.libsonnet | 26 + .../_gen/storage/v1/volumeError.libsonnet | 10 + .../storage/v1/volumeNodeResources.libsonnet | 8 + .../1.19/_gen/storage/v1alpha1/main.libsonnet | 9 + .../v1alpha1/volumeAttachment.libsonnet | 479 + .../v1alpha1/volumeAttachmentSource.libsonnet | 419 + .../v1alpha1/volumeAttachmentSpec.libsonnet | 426 + .../v1alpha1/volumeAttachmentStatus.libsonnet | 26 + .../storage/v1alpha1/volumeError.libsonnet | 10 + .../_gen/storage/v1beta1/csiDriver.libsonnet | 71 + .../storage/v1beta1/csiDriverSpec.libsonnet | 18 + .../_gen/storage/v1beta1/csiNode.libsonnet | 63 + .../storage/v1beta1/csiNodeDriver.libsonnet | 19 + .../storage/v1beta1/csiNodeSpec.libsonnet | 10 + .../1.19/_gen/storage/v1beta1/main.libsonnet | 16 + .../storage/v1beta1/storageClass.libsonnet | 76 + .../v1beta1/volumeAttachment.libsonnet | 479 + .../v1beta1/volumeAttachmentSource.libsonnet | 419 + .../v1beta1/volumeAttachmentSpec.libsonnet | 426 + .../v1beta1/volumeAttachmentStatus.libsonnet | 26 + .../storage/v1beta1/volumeError.libsonnet | 10 + .../v1beta1/volumeNodeResources.libsonnet | 8 + .../k8s-libsonnet/1.19/gen.libsonnet | 25 + .../k8s-libsonnet/1.19/main.libsonnet | 1 + .../ksonnet-lib/ksonnet.beta.4/k.libsonnet | 123 - .../ksonnet-lib/ksonnet.beta.4/k8s.libsonnet | 30584 ---------------- example/tk/vendor/ksonnet.beta.4 | 1 - .../jsonnet/single-binary/tempo.libsonnet | 15 +- 560 files changed, 27335 insertions(+), 30765 deletions(-) create mode 120000 example/tk/vendor/1.19 create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/apps.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/batch.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/core.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/list.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/mapContainers.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/rbac.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/volumeMounts.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhook.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhookConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/ruleWithOperations.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/serviceReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhook.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhookConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/webhookClientConfig.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhook.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhookConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/ruleWithOperations.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/serviceReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhook.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhookConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/webhookClientConfig.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiService.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/serviceReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiService.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/serviceReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/controllerRevision.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSet.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetUpdateStrategy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deployment.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStrategy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSet.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDaemonSet.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDeployment.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateStatefulSetStrategy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSet.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetUpdateStrategy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/boundObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequest.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/userInfo.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/userInfo.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/localSubjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceAttributes.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceAttributes.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectRulesReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/localSubjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceAttributes.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceAttributes.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReview.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectRulesReviewStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/crossVersionObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscaler.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scale.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/crossVersionObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscaler.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/crossVersionObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscaler.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerBehavior.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingPolicy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingRules.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricIdentifier.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricTarget.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricValueStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/job.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJob.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/jobTemplateSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJob.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/jobTemplateSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequest.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequest.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/lease.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/leaseSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/lease.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/leaseSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/affinity.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/attachedVolume.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/awsElasticBlockStoreVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureDiskVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFilePersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFileVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/binding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/capabilities.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/clientIPConfig.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMap.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapEnvSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapKeySelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapNodeConfigSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapProjection.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/container.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerImage.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerPort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerState.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateRunning.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateTerminated.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateWaiting.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/daemonEndpoint.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIProjection.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeFile.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/emptyDirVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointAddress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointPort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointSubset.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpoints.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envFromSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVar.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVarSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralContainer.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/event.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSeries.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/execAction.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/fcVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flockerVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gcePersistentDiskVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gitRepoVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/handler.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostAlias.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostPathVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpGetAction.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpHeader.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/keyToPath.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/lifecycle.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRange.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeItem.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerIngress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespace.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nfsVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/node.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAddress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAffinity.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeDaemonEndpoints.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorRequirement.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorTerm.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSystemInfo.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectFieldSelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolume.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaim.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimTemplate.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/photonPersistentDiskVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/pod.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinity.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinityTerm.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAntiAffinity.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfig.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfigOption.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podIP.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podReadinessGate.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSecurityContext.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplate.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplateSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/portworxVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/preferredSchedulingTerm.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/probe.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/projectedVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/quobyteVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationController.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceFieldSelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuota.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceRequirements.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopeSelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopedResourceSelectorRequirement.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seLinuxOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seccompProfile.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secret.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretEnvSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretKeySelector.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretProjection.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/securityContext.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/service.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccount.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccountTokenProjection.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/servicePort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sessionAffinityConfig.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSPersistentVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sysctl.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/taint.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/tcpSocketAction.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/toleration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorLabelRequirement.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorTerm.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySpreadConstraint.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/typedLocalObjectReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volume.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeDevice.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeMount.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeNodeAffinity.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeProjection.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/vsphereVirtualDiskVolumeSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/weightedPodAffinityTerm.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/windowsSecurityContextOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpoint.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointConditions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointPort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointSlice.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/event.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/eventSeries.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/event.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/eventSeries.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressPath.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressRuleValue.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressBackend.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressTLS.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowDistinguisherMethod.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchema.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/groupSubject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitResponse.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitedPriorityLevelConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/nonResourcePolicyRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/policyRulesWithSubjects.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationCondition.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationReference.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/queuingConfiguration.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/resourcePolicyRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/serviceAccountSubject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/subject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/userSubject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressPath.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressRuleValue.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressBackend.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClassSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressServiceBackend.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressTLS.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ipBlock.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyEgressRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyIngressRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPeer.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicySpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/serviceBackendPort.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressPath.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressRuleValue.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingress.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressBackend.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClassSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressTLS.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/overhead.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClassSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/scheduling.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/overhead.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/runtimeClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/scheduling.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedCSIDriver.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedFlexVolume.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedHostPath.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/eviction.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/fsGroupStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/hostPortRange.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/idRange.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudget.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicy.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicySpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsGroupStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsUserStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runtimeClassStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/seLinuxStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/supplementalGroupsStrategyOptions.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/aggregationRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRole.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRoleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/policyRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/role.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleRef.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/subject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/aggregationRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRole.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRoleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/policyRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/role.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleRef.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/subject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/aggregationRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRole.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRoleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/policyRule.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/role.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleBinding.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleRef.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/subject.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/priorityClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/priorityClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/priorityClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPreset.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPresetSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriver.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriverSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNode.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeDriver.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/storageClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachment.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeError.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeNodeResources.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachment.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeError.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriver.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriverSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNode.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeDriver.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/main.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/storageClass.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachment.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSource.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSpec.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentStatus.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeError.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeNodeResources.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/gen.libsonnet create mode 100644 example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet delete mode 100644 example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k.libsonnet delete mode 100644 example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k8s.libsonnet delete mode 120000 example/tk/vendor/ksonnet.beta.4 diff --git a/CHANGELOG.md b/CHANGELOG.md index d7c3fa5ccd1..5095fac62ba 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -51,6 +51,7 @@ * [ENHANCEMENT] Support global denylist and per-tenant allowlist of tags for search data. [#960](https://github.com/grafana/tempo/pull/960) (@annanay25) * [ENHANCEMENT] Add `search_query_timeout` to Querier config. [#984](https://github.com/grafana/tempo/pull/984) (@kvrhdn) * [ENHANCEMENT] Jsonnet: add `$._config.memcached.memory_limit_mb` [#987](https://github.com/grafana/tempo/pull/987) (@kvrhdn) +* [ENHANCEMENT] Upgrade jsonnet-libs to 1.19 and update tk examples [#](https://github.com/grafana/tempo/pull/) (@mapno) * [BUGFIX] Update port spec for GCS docker-compose example [#869](https://github.com/grafana/tempo/pull/869) (@zalegrala) * [BUGFIX] Fix "magic number" errors and other block mishandling when an ingester forcefully shuts down [#937](https://github.com/grafana/tempo/issues/937) (@mdisibio) * [BUGFIX] Fix compactor memory leak [#806](https://github.com/grafana/tempo/pull/806) (@mdisibio) diff --git a/example/tk/jsonnetfile.json b/example/tk/jsonnetfile.json index 87a5531f518..b855b76709e 100644 --- a/example/tk/jsonnetfile.json +++ b/example/tk/jsonnetfile.json @@ -22,11 +22,11 @@ { "source": { "git": { - "remote": "https://github.com/ksonnet/ksonnet-lib.git", - "subdir": "ksonnet.beta.4" + "remote": "https://github.com/jsonnet-libs/k8s-libsonnet.git", + "subdir": "1.19" } }, - "version": "master" + "version": "main" } ], "legacyImports": true diff --git a/example/tk/jsonnetfile.lock.json b/example/tk/jsonnetfile.lock.json index f73f2097a18..6d396c292d9 100644 --- a/example/tk/jsonnetfile.lock.json +++ b/example/tk/jsonnetfile.lock.json @@ -8,8 +8,8 @@ "subdir": "ksonnet-util" } }, - "version": "eae352a28812274df34ccee8a7edec3a97a1ef95", - "sum": "sI6Hgng7yR8fkgHqJ10e3rQagVhpHhumAKqqzbOZUSM=" + "version": "9631fe778e87213bed195fcdc8d79da4eecb7431", + "sum": "OxgtIWL4hjvG0xkMwUzZ7Yjs52zUhLhaVQpwHCbqf8A=" }, { "source": { @@ -18,18 +18,18 @@ "subdir": "memcached" } }, - "version": "2829b0f8e18ec87c3a451fa01fcd11b6508b25b2", - "sum": "AIspZ151p0qkxVc9tuoAEYNrkazV6QncWWsIsarK9GE=" + "version": "9631fe778e87213bed195fcdc8d79da4eecb7431", + "sum": "dTOeEux3t9bYSqP2L/uCuLo/wUDpCKH4w+4OD9fePUk=" }, { "source": { "git": { - "remote": "https://github.com/ksonnet/ksonnet-lib.git", - "subdir": "ksonnet.beta.4" + "remote": "https://github.com/jsonnet-libs/k8s-libsonnet.git", + "subdir": "1.19" } }, - "version": "0d2f82676817bbf9e4acf6495b2090205f323b9f", - "sum": "ur22hPQq0JAPBxm8hNMcwjumj4MkozDwOKiGZvVMYh4=" + "version": "f47562c4e3bfaf46a5a5bc6a9e1e76a4fc76c18e", + "sum": "m/53207lTQJGAA9H2D7SrM+qIMJxHxXPzlsx3GuyYLg=" } ], "legacyImports": false diff --git a/example/tk/lib/grafana/main.libsonnet b/example/tk/lib/grafana/main.libsonnet index ab50699f573..893d87b8add 100644 --- a/example/tk/lib/grafana/main.libsonnet +++ b/example/tk/lib/grafana/main.libsonnet @@ -55,6 +55,10 @@ grafana_service: k.util.serviceFor($.grafana_deployment) + service.mixin.spec.withPortsMixin([ - servicePort.withName('http').withPort(3000).withTargetPort(3000), + servicePort.newNamed( + name='http', + port=3000, + targetPort=3000, + ), ]), } diff --git a/example/tk/lib/k.libsonnet b/example/tk/lib/k.libsonnet index 03d16e05222..f689e390078 100644 --- a/example/tk/lib/k.libsonnet +++ b/example/tk/lib/k.libsonnet @@ -1 +1 @@ -import "ksonnet.beta.4/k.libsonnet" \ No newline at end of file +import 'github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet' \ No newline at end of file diff --git a/example/tk/lib/minio/minio.libsonnet b/example/tk/lib/minio/minio.libsonnet index 4fa00b50c9a..28647c566b5 100644 --- a/example/tk/lib/minio/minio.libsonnet +++ b/example/tk/lib/minio/minio.libsonnet @@ -13,7 +13,7 @@ container.withCommand([ 'sh', '-euc', - 'mkdir -p /data/tempo && /usr/bin/minio server /data', + 'mkdir -p /data/tempo && minio server /data', ]) + container.withEnvMap({ MINIO_ACCESS_KEY: 'tempo', diff --git a/example/tk/tempo-microservices/main.jsonnet b/example/tk/tempo-microservices/main.jsonnet index 4d1f752a406..e788b6b0b77 100644 --- a/example/tk/tempo-microservices/main.jsonnet +++ b/example/tk/tempo-microservices/main.jsonnet @@ -104,19 +104,21 @@ minio + grafana + load + tempo { } }, - local ingress = k.networking.v1beta1.ingress, + local ingress = k.networking.v1.ingress, + local rule = k.networking.v1.ingressRule, + local path = k.networking.v1.httpIngressPath, ingress: - ingress.new() + + ingress.new('ingress') + ingress.mixin.metadata - .withName('ingress') - .withAnnotations({ + .withAnnotationsMixin({ 'ingress.kubernetes.io/ssl-redirect': 'false' }) + ingress.mixin.spec.withRules( - ingress.mixin.specType.rulesType.mixin.http.withPaths( - ingress.mixin.spec.rulesType.mixin.httpType.pathsType.withPath('/') + - ingress.mixin.specType.mixin.backend.withServiceName('grafana') + - ingress.mixin.specType.mixin.backend.withServicePort(3000) - ), + rule.http.withPaths([ + path.withPath('/') + + path.withPathType('ImplementationSpecific') + + path.backend.service.withName('grafana') + + path.backend.service.port.withNumber(3000) + ]), ), } diff --git a/example/tk/tempo-single-binary/main.jsonnet b/example/tk/tempo-single-binary/main.jsonnet index 5483a294556..d7164d6ca53 100644 --- a/example/tk/tempo-single-binary/main.jsonnet +++ b/example/tk/tempo-single-binary/main.jsonnet @@ -28,19 +28,21 @@ grafana + load + tempo { containerPort.new('jaeger-http', 14268), ]), - local ingress = k.networking.v1beta1.ingress, + local ingress = k.networking.v1.ingress, + local rule = k.networking.v1.ingressRule, + local path = k.networking.v1.httpIngressPath, ingress: - ingress.new() + + ingress.new('ingress') + ingress.mixin.metadata - .withName('ingress') - .withAnnotations({ + .withAnnotationsMixin({ 'ingress.kubernetes.io/ssl-redirect': 'false' }) + ingress.mixin.spec.withRules( - ingress.mixin.specType.rulesType.mixin.http.withPaths( - ingress.mixin.spec.rulesType.mixin.httpType.pathsType.withPath('/') + - ingress.mixin.specType.mixin.backend.withServiceName('grafana') + - ingress.mixin.specType.mixin.backend.withServicePort(3000) - ), + rule.http.withPaths([ + path.withPath('/') + + path.withPathType('ImplementationSpecific') + + path.backend.service.withName('grafana') + + path.backend.service.port.withNumber(3000) + ]), ), } diff --git a/example/tk/vendor/1.19 b/example/tk/vendor/1.19 new file mode 120000 index 00000000000..3b695724e53 --- /dev/null +++ b/example/tk/vendor/1.19 @@ -0,0 +1 @@ +github.com/jsonnet-libs/k8s-libsonnet/1.19 \ No newline at end of file diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/k-compat.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/k-compat.libsonnet index 65b5b6e26ba..389dcab7375 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/k-compat.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/k-compat.libsonnet @@ -1,4 +1,4 @@ -// k-compat.libsonnet provides a compatibility layer between k8s-alpha and ksonnet-lib. As ksonnet-lib has been +// k-compat.libsonnet provides a compatibility layer between k8s-libsonnet and ksonnet-lib. As ksonnet-lib has been // abandoned, we consider it deprecated. This layer will generate a deprecation warning to those that still use it. local k = import 'k.libsonnet'; @@ -7,7 +7,7 @@ k if std.objectHas(k, '__ksonnet') then std.trace( - 'Deprecated: ksonnet-lib has been abandoned, please consider using https://github.com/jsonnet-libs/k8s-alpha.', + 'Deprecated: ksonnet-lib has been abandoned, please consider using https://github.com/jsonnet-libs/k8s-libsonnet.', (import 'legacy-types.libsonnet') + (import 'legacy-custom.libsonnet') + (import 'legacy-noname.libsonnet')({ diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-custom.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-custom.libsonnet index e81e6a85f27..2d42997ab8f 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-custom.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-custom.libsonnet @@ -1,4 +1,4 @@ -// legacy-custom.libsonnet retrofits k8s-alpha functionality into ksonnet-lib +// legacy-custom.libsonnet retrofits k8s-libsonnet functionality into ksonnet-lib { core+: { v1+: { diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-noname.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-noname.libsonnet index 388a10b98ed..9529244f299 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-noname.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-noname.libsonnet @@ -1,4 +1,4 @@ -// legacy-noname.libsonnet provides two-way compatibility, in k8s-alpha many new() functions have a mandatory name +// legacy-noname.libsonnet provides two-way compatibility, in k8s-libsonnet many new() functions have a mandatory name // argument while they are absent in ksonnet-lib. `noNewEmptyNameMixin` allows us to make the argument optional in // either situation. function(noNewEmptyNameMixin) { diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-subtypes.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-subtypes.libsonnet index 0beddfcd7a6..286928648ea 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-subtypes.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/legacy-subtypes.libsonnet @@ -1,6 +1,6 @@ // legacy-subtypes.libsonnet makes first-class entities available as subtypes like in ksonnet-lib. // It also makes the empty new() functions and camelCased functions available. -// This is largely based on kausal-shim.libsonnet from k8s-alpha. +// This is largely based on kausal-shim.libsonnet from k8s-libsonnet. { core+: { v1+: { container+: { diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/util.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/util.libsonnet index c9458466ae1..d97f7d5db32 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/util.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/ksonnet-util/util.libsonnet @@ -84,7 +84,7 @@ local util(k) = { role.mixin.metadata.withNamespace(namespace) + role.withRules(rules), - cluster_role_binding: + role_binding: roleBinding.new() + roleBinding.mixin.metadata.withName(name) + roleBinding.mixin.metadata.withNamespace(namespace) + @@ -155,6 +155,21 @@ local util(k) = { volume.fromHostPath(name, hostPath), ]), + pvcVolumeMount(pvcName, path, readOnly=false, volumeMountMixin={}):: + local container = k.core.v1.container, + deployment = k.apps.v1.deployment, + volumeMount = k.core.v1.volumeMount, + volume = k.core.v1.volume, + addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(pvcName, path, readOnly=readOnly) + + volumeMountMixin, + ); + + deployment.mapContainers(addMount) + + deployment.mixin.spec.template.spec.withVolumesMixin([ + volume.fromPersistentVolumeClaim(pvcName, pvcName), + ]), + secretVolumeMount(name, path, defaultMode=256, volumeMountMixin={}):: local container = k.core.v1.container, deployment = k.apps.v1.deployment, diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/README.md b/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/README.md index 607deaabaa3..4922e216aa7 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/README.md +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/README.md @@ -2,25 +2,28 @@ A set of extensible configs for running Memcached on Kubernetes. -To use this libary, install [Tanka](https://tanka.dev/) and [Jsonnet Bundler](https://tanka.dev/install#jsonnet-bundler). +To use this libary, install [Tanka] and [Jsonnet Bundler]. + +[Tanka]: https://tanka.dev/ +[Jsonnet Bundler]: https://tanka.dev/install#jsonnet-bundler In your config repo, if you don't have a Tanka application, make a new one (will copy credentials from current context): -``` -$ mkdir -$ cd -$ tk init +```bash +app=your_app_name +mkdir -p $app && cd $app +tk init ``` Then you can install the library with: -``` +```bash jb install github.com/grafana/jsonnet-libs/memcached ``` Assuming you want to run in the default namespace ('environment' in Tanka parlance), add the following to the file `environments/default/main.jsonnet`: -``` +```jsonnet local memcached = import "memcached/memcached.libsonnet"; memcached + { @@ -33,19 +36,20 @@ memcached + { } ``` -- Apply your config: +Apply your config: +```shell +tk apply default ``` -$ tk apply default -``` -# Customising and Extending. + +## Customising and Extending The choice of Tanka for configuring these jobs was intentional; it allows users to easily override setting in these configurations to suit their needs, without having to fork or modify this library. For instance, to override the resource requests and limits for the Consul container, you would: -``` +```jsonnet local memcached = import "memcached/memcached.libsonnet"; memcached + { @@ -65,7 +69,7 @@ memcached + { We sometimes specify config options in a `_config` dict; there are two situations under which we do this: -- When you must provide a value for the parameter (such as `namesapce`). +- When you must provide a value for the parameter (such as `namespace`). - When the parameter get referenced in multiple places, and overriding it using the technique above would be cumbersome and error prone (such as with `cluster_dns_suffix`). diff --git a/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/memcached.libsonnet b/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/memcached.libsonnet index 021ee6d2f25..3d2dda59bb9 100644 --- a/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/memcached.libsonnet +++ b/example/tk/vendor/github.com/grafana/jsonnet-libs/memcached/memcached.libsonnet @@ -25,10 +25,11 @@ k { overprovision_factor:: 1.2, cpu_limits:: '3', connection_limit:: 1024, + memory_request_overhead_mb:: 100, memory_request_bytes:: - std.ceil((self.memory_limit_mb * self.overprovision_factor) + 100) * 1024 * 1024, + std.ceil((self.memory_limit_mb * self.overprovision_factor) + self.memory_request_overhead_mb) * 1024 * 1024, memory_limits_bytes:: - self.memory_limit_mb * 1.5 * 1024 * 1024, + std.max(self.memory_limit_mb * 1.5 * 1024 * 1024, self.memory_request_bytes), local container = $.core.v1.container, local containerPort = $.core.v1.containerPort, diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/apps.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/apps.libsonnet new file mode 100644 index 00000000000..27a4b615704 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/apps.libsonnet @@ -0,0 +1,83 @@ +local d = import 'doc-util/main.libsonnet'; + +local patch = { + daemonSet+: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('containers', d.T.array), + d.arg('podLabels', d.T.object, {}), + ]), + new( + name, + containers=[], + podLabels={} + ):: + local labels = { name: name } + podLabels; + super.new(name) + + super.spec.template.spec.withContainers(containers) + + super.spec.template.metadata.withLabels(labels) + + super.spec.selector.withMatchLabels(labels), + }, + deployment+: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('replicas', d.T.int, 1), + d.arg('containers', d.T.array), + d.arg('podLabels', d.T.object, {}), + ]), + new( + name, + replicas=1, + containers=error 'containers unset', + podLabels={}, + ):: + local labels = { name: name } + podLabels; + super.new(name) + + super.spec.withReplicas(replicas) + + super.spec.template.spec.withContainers(containers) + + super.spec.template.metadata.withLabels(labels) + + super.spec.selector.withMatchLabels(labels), + }, + + statefulSet+: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('replicas', d.T.int, 1), + d.arg('containers', d.T.array), + d.arg('volumeClaims', d.T.array, []), + d.arg('podLabels', d.T.object, {}), + ]), + new( + name, + replicas=1, + containers=error 'containers unset', + volumeClaims=[], + podLabels={}, + ):: + local labels = { name: name } + podLabels; + super.new(name) + + super.spec.withReplicas(replicas) + + super.spec.template.spec.withContainers(containers) + + super.spec.template.metadata.withLabels(labels) + + super.spec.selector.withMatchLabels(labels) + + // remove volumeClaimTemplates if empty + // (otherwise it will create a diff all the time) + + ( + if std.length(volumeClaims) > 0 + then super.spec.withVolumeClaimTemplates(volumeClaims) + else {} + ), + }, +}; + +{ + extensions+: { // old, remove asap + v1beta1+: patch, + }, + apps+: { + v1+: patch, + v1beta1+: patch, + v1beta2+: patch, + }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/batch.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/batch.libsonnet new file mode 100644 index 00000000000..c4ccdfc3de8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/batch.libsonnet @@ -0,0 +1,27 @@ +local d = import 'doc-util/main.libsonnet'; + +local patch = { + cronJob+: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('schedule', d.T.string), + d.arg('containers', d.T.array), + ]), + new( + name, + schedule='', + containers=[] + ):: + super.new(name) + + super.spec.withSchedule(schedule) + + super.spec.jobTemplate.spec.template.spec.withContainers(containers) + + super.spec.jobTemplate.spec.template.metadata.withLabels({ name: name }), + }, +}; + +{ + batch+: { + v1beta1+: patch, + v2alpha1+: patch, + }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/core.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/core.libsonnet new file mode 100644 index 00000000000..b342d37e2b2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/core.libsonnet @@ -0,0 +1,214 @@ +local d = import 'doc-util/main.libsonnet'; + +{ + core+: { + v1+: { + configMap+: { + local withData(data) = if data != {} then super.withData(data) else {}, + withData:: withData, + + local withDataMixin(data) = if data != {} then super.withDataMixin(data) else {}, + withDataMixin:: withDataMixin, + + '#new': d.fn('new creates a new `ConfigMap` of given `name` and `data`', [d.arg('name', d.T.string), d.arg('data', d.T.object)]), + new(name, data={}):: + super.new(name) + + super.metadata.withName(name) + + withData(data), + }, + + container+: { + '#new': d.fn('new returns a new `container` of given `name` and `image`', [d.arg('name', d.T.string), d.arg('image', d.T.string)]), + new(name, image):: super.withName(name) + super.withImage(image), + + withEnvMixin(env):: + // if an envvar has an empty value ("") we want to remove that property + // because k8s will remove that and then it would always + // show up as a difference. + local removeEmptyValue(obj) = + if std.objectHas(obj, 'value') && std.length(obj.value) == 0 then + { + [k]: obj[k] + for k in std.objectFields(obj) + if k != 'value' + } + else + obj; + super.withEnvMixin([ + removeEmptyValue(envvar) + for envvar in env + ]), + + '#withEnvMap': d.fn( + '`withEnvMap` works like `withEnvMixin` but accepts a key/value map, this map is converted a list of core.v1.envVar(key, value)`', + [d.arg('env', d.T.object)] + ), + withEnvMap(env):: + self.withEnvMixin([ + $.core.v1.envVar.new(k, env[k]) + for k in std.objectFields(env) + ]), + }, + + containerPort+: { + // using a local here to re-use new, because it is lexically scoped, + // while `self` is not + local new(containerPort) = super.withContainerPort(containerPort), + local newNamed(containerPort, name) = new(containerPort) + super.withName(name), + '#new': d.fn('new returns a new `containerPort`', [d.arg('containerPort', d.T.int)]), + new:: new, + '#newNamed': d.fn('newNamed works like `new`, but also sets the `name`', [d.arg('containerPort', d.T.int), d.arg('name', d.T.string)]), + newNamed:: newNamed, + '#newUDP': d.fn('newUDP works like `new`, but also sets protocal to UDP', [d.arg('containerPort', d.T.int)]), + newUDP(containerPort):: new(containerPort) + super.withProtocol('UDP'), + '#newNamedUDP': d.fn('newNamedUDP works like `newNamed`, but also sets protocal to UDP', [d.arg('containerPort', d.T.int), d.arg('name', d.T.string)]), + newNamedUDP(containerPort, name):: newNamed(containerPort, name) + super.withProtocol('UDP'), + }, + + envVar+: { + '#new': d.fn('new returns a new `envVar` of given `name` and `value`', [d.arg('name', d.T.string), d.arg('value', d.T.string)]), + new(name, value):: super.withName(name) + super.withValue(value), + + '#fromSecretRef': d.fn('fromSecretRef creates a `envVar` from a secret reference', [ + d.arg('name', d.T.string), + d.arg('secretRefName', d.T.string), + d.arg('secretRefKey', d.T.string), + ]), + fromSecretRef(name, secretRefName, secretRefKey):: + super.withName(name) + + super.valueFrom.secretKeyRef.withName(secretRefName) + + super.valueFrom.secretKeyRef.withKey(secretRefKey), + + '#fromFieldPath': d.fn('fromFieldPath creates a `envVar` from a field path', [ + d.arg('name', d.T.string), + d.arg('fieldPath', d.T.string), + ]), + fromFieldPath(name, fieldPath):: + super.withName(name) + + super.valueFrom.fieldRef.withFieldPath(fieldPath), + }, + + keyToPath+:: { + '#new': d.fn('new creates a new `keyToPath`', [d.arg('key', d.T.string), d.arg('path', d.T.string)]), + new(key, path):: super.withKey(key) + super.withPath(path), + }, + + persistentVolume+: { + '#new':: d.fn(help='new returns an instance of Persistentvolume', args=[d.arg(name='name', type=d.T.string)]), + new(name=''): { + apiVersion: 'v1', + kind: 'PersistentVolume', + } + ( + if name != '' + then self.metadata.withName(name=name) + else {} + ), + }, + + secret+:: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('data', d.T.object), + d.arg('type', d.T.string, 'Opaque'), + ]), + new(name, data, type='Opaque'):: + super.new(name) + + super.withData(data) + + super.withType(type), + }, + + service+:: { + '#new'+: d.func.withArgs([ + d.arg('name', d.T.string), + d.arg('selector', d.T.object), + d.arg('ports', d.T.array), + ]), + new(name, selector, ports):: + super.new(name) + + super.spec.withSelector(selector) + + super.spec.withPorts(ports), + }, + + servicePort+:: { + local new(port, targetPort) = super.withPort(port) + super.withTargetPort(targetPort), + '#new': d.fn('new returns a new `servicePort`', [ + d.arg('port', d.T.int), + d.arg('targetPort', d.T.any), + ]), + new:: new, + + '#newNamed': d.fn('newNamed works like `new`, but also sets the `name`', [ + d.arg('name', d.T.string), + d.arg('port', d.T.int), + d.arg('targetPort', d.T.any), + ]), + newNamed(name, port, targetPort):: + new(port, targetPort) + super.withName(name), + }, + + volume+:: { + '#fromConfigMap': d.fn('Creates a new volume from a `ConfigMap`', [ + d.arg('name', d.T.string), + d.arg('configMapName', d.T.string), + d.arg('configMapItems', d.T.array), + ]), + fromConfigMap(name, configMapName, configMapItems=[]):: + super.withName(name) + + super.configMap.withName(configMapName) + + ( + if configMapItems != [] + then super.configMap.withItems(configMapItems) + else {} + ), + + '#fromEmptyDir': d.fn('Creates a new volume of type `emptyDir`', [ + d.arg('name', d.T.string), + d.arg('emptyDir', d.T.object, {}), + ]), + fromEmptyDir(name, emptyDir={}):: + super.withName(name) + { emptyDir: emptyDir }, + + '#fromPersistentVolumeClaim': d.fn('Creates a new volume using a `PersistentVolumeClaim`.', [ + d.arg('name', d.T.string), + d.arg('claimName', d.T.string), + ]), + fromPersistentVolumeClaim(name, claimName='', emptyDir=''):: + // Note: emptyDir is inherited from ksonnet-lib, this provides backwards compatibility + local claim = + if (claimName == '' && emptyDir != '') + then emptyDir + else + if claimName == '' + then error 'claimName not set' + else claimName; + super.withName(name) + super.persistentVolumeClaim.withClaimName(claim), + + '#fromHostPath': d.fn('Creates a new volume using a `hostPath`', [ + d.arg('name', d.T.string), + d.arg('hostPath', d.T.string), + ]), + fromHostPath(name, hostPath):: + super.withName(name) + super.hostPath.withPath(hostPath), + + '#fromSecret': d.fn('Creates a new volume from a `Secret`', [ + d.arg('name', d.T.string), + d.arg('secretName', d.T.string), + ]), + fromSecret(name, secretName):: + super.withName(name) + super.secret.withSecretName(secretName), + }, + + volumeMount+:: { + '#new': d.fn('new creates a new `volumeMount`', [ + d.arg('name', d.T.string), + d.arg('mountPath', d.T.string), + d.arg('readOnly', d.T.bool), + ]), + new(name, mountPath, readOnly=false):: + super.withName(name) + + super.withMountPath(mountPath) + + (if readOnly then self.withReadOnly(readOnly) else {}), + }, + }, + }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/list.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/list.libsonnet new file mode 100644 index 00000000000..44a151cf9bd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/list.libsonnet @@ -0,0 +1,27 @@ +local d = import 'doc-util/main.libsonnet'; + +{ + core+: { + v1+: { + list: { + '#':: d.pkg(name='list', url='', help='List represents a generic list of items.'), + '#new': d.fn( + '`new` returns an instance of List.', + [d.arg('items', d.T.array)] + ), + new(items):: { + apiVersion: 'v1', + kind: 'List', + } + self.withItems(items), + '#withItems': d.fn( + '`withItems` List of items to populate the items in a list.', + [d.arg('items', d.T.array)] + ), + withItems(items):: + if std.isArray(v=items) + then { items+: items } + else { items+: [items] }, + }, + }, + }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/mapContainers.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/mapContainers.libsonnet new file mode 100644 index 00000000000..66d4e44892a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/mapContainers.libsonnet @@ -0,0 +1,74 @@ +local d = import 'doc-util/main.libsonnet'; + +local patch = { + '#mapContainers': d.fn( + ||| + `mapContainers` applies the function f to each container. + It works exactly as `std.map`, but on the containers of this object. + + **Signature of `f`**: + ```ts + f(container: Object) Object + ``` + |||, + [d.arg('f', d.T.func)] + ), + mapContainers(f):: { + local podContainers = super.spec.template.spec.containers, + spec+: { + template+: { + spec+: { + containers: std.map(f, podContainers), + }, + }, + }, + }, + + '#mapContainersWithName': d.fn('`mapContainersWithName` is like `mapContainers`, but only applies to those containers in the `names` array', + [d.arg('names', d.T.array), d.arg('f', d.T.func)]), + mapContainersWithName(names, f):: + local nameSet = if std.type(names) == 'array' then std.set(names) else std.set([names]); + local inNameSet(name) = std.length(std.setInter(nameSet, std.set([name]))) > 0; + + self.mapContainers(function(c) if std.objectHas(c, 'name') && inNameSet(c.name) then f(c) else c), +}; + +// batch.job and batch.cronJob have the podSpec at a different location +local cronPatch = patch { + mapContainers(f):: { + local podContainers = super.spec.jobTemplate.spec.template.spec.containers, + spec+: { + jobTemplate+: { + spec+: { + template+: { + spec+: { + containers: std.map(f, podContainers), + }, + }, + }, + }, + }, + }, +}; + +{ + core+: { v1+: { + pod+: patch, + podTemplate+: patch, + replicationController+: patch, + } }, + batch+: { + v1+: { + job+: patch, + }, + v1beta1+: { + cronJob+: cronPatch, + }, + }, + apps+: { v1+: { + daemonSet+: patch, + deployment+: patch, + replicaSet+: patch, + statefulSet+: patch, + } }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/rbac.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/rbac.libsonnet new file mode 100644 index 00000000000..69327fb6116 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/rbac.libsonnet @@ -0,0 +1,43 @@ +local d = import 'doc-util/main.libsonnet'; + +local bindRoleDoc = d.fn( + '`bindRole` returns a roleRef for a Role or ClusterRole object.', + [d.arg('role', d.T.object)] +); + +local bindRole(role) = { + roleRef: { + name: role.metadata.name, + kind: role.kind, + apiGroup: std.split(role.apiVersion, '/')[0], + }, +}; + +local patch = { + roleBinding+: { + '#bindRole': bindRoleDoc, + bindRole(role):: bindRole(role), + }, + clusterRoleBinding+: { + '#bindRole': bindRoleDoc, + bindRole(role):: bindRole(role), + }, + subject+: { + '#fromServiceAccount': d.fn( + '`fromServiceAccount` returns a subject for a service account.', + [d.arg('service_account', d.T.object)] + ), + fromServiceAccount(service_account):: + super.withKind('ServiceAccount') + + super.withName(service_account.metadata.name) + + super.withNamespace(service_account.metadata.namespace), + }, +}; + +{ + rbac+: { + v1+: patch, + v1alpha1+: patch, + v1beta1+: patch, + }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/volumeMounts.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/volumeMounts.libsonnet new file mode 100644 index 00000000000..2f3e74aaaee --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_custom/volumeMounts.libsonnet @@ -0,0 +1,169 @@ +local d = import 'doc-util/main.libsonnet'; + +{ + local container = $.core.v1.container, + local volumeMount = $.core.v1.volumeMount, + local volume = $.core.v1.volume, + + local patch = { + local volumeMountDescription = + ||| + This helper function can be augmented with a `volumeMountsMixin. For example, + passing "k.core.v1.volumeMount.withSubPath(subpath)" will result in a subpath + mixin. + |||, + + + '#configVolumeMount': d.fn( + '`configVolumeMount` mounts a ConfigMap by `name` into all container on `path`.' + + volumeMountDescription, + [ + d.arg('name', d.T.string), + d.arg('path', d.T.string), + d.arg('volumeMountMixin', d.T.object), + ] + ), + configVolumeMount(name, path, volumeMountMixin={}):: + local addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromConfigMap(name, name), + ]), + + + '#configMapVolumeMount': d.fn( + ||| + `configMapVolumeMount` mounts a `configMap` into all container on `path`. It will + also add an annotation hash to ensure the pods are re-deployed when the config map + changes. + ||| + + volumeMountDescription, + [ + d.arg('configMap', d.T.object), + d.arg('path', d.T.string), + d.arg('volumeMountMixin', d.T.object), + ] + ), + configMapVolumeMount(configMap, path, volumeMountMixin={}):: + local name = configMap.metadata.name, + hash = std.md5(std.toString(configMap)), + addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromConfigMap(name, name), + ]) + + super.spec.template.metadata.withAnnotationsMixin({ + ['%s-hash' % name]: hash, + }), + + + '#hostVolumeMount': d.fn( + '`hostVolumeMount` mounts a `hostPath` into all container on `path`.' + + volumeMountDescription, + [ + d.arg('name', d.T.string), + d.arg('hostPath', d.T.string), + d.arg('path', d.T.string), + d.arg('readOnly', d.T.bool), + d.arg('volumeMountMixin', d.T.object), + ] + ), + hostVolumeMount(name, hostPath, path, readOnly=false, volumeMountMixin={}):: + local addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path, readOnly=readOnly) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromHostPath(name, hostPath), + ]), + + + '#pvcVolumeMount': d.fn( + '`hostVolumeMount` mounts a PersistentVolumeClaim by `name` into all container on `path`.' + + volumeMountDescription, + [ + d.arg('name', d.T.string), + d.arg('path', d.T.string), + d.arg('readOnly', d.T.bool), + d.arg('volumeMountMixin', d.T.object), + ] + ), + pvcVolumeMount(name, path, readOnly=false, volumeMountMixin={}):: + local addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path, readOnly=readOnly) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromPersistentVolumeClaim(name, name), + ]), + + + '#secretVolumeMount': d.fn( + '`secretVolumeMount` mounts a Secret by `name` into all container on `path`.' + + volumeMountDescription, + [ + d.arg('name', d.T.string), + d.arg('path', d.T.string), + d.arg('defaultMode', d.T.string), + d.arg('volumeMountMixin', d.T.object), + ] + ), + secretVolumeMount(name, path, defaultMode=256, volumeMountMixin={}):: + local addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromSecret(name, secretName=name) + + volume.secret.withDefaultMode(defaultMode), + ]), + + + '#emptyVolumeMount': d.fn( + '`emptyVolumeMount` mounts empty volume by `name` into all container on `path`.' + + volumeMountDescription, + [ + d.arg('name', d.T.string), + d.arg('path', d.T.string), + d.arg('volumeMountMixin', d.T.object), + d.arg('volumeMixin', d.T.object), + ] + ), + emptyVolumeMount(name, path, volumeMountMixin={}, volumeMixin={}):: + local addMount(c) = c + container.withVolumeMountsMixin( + volumeMount.new(name, path) + + volumeMountMixin, + ); + + super.mapContainers(addMount) + + super.spec.template.spec.withVolumesMixin([ + volume.fromEmptyDir(name) + volumeMixin, + ]), + }, + + batch+: { + v1+: { + job+: patch, + }, + }, + apps+: { v1+: { + daemonSet+: patch, + deployment+: patch, + replicaSet+: patch, + statefulSet+: patch, + } }, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/main.libsonnet new file mode 100644 index 00000000000..5c9c107733f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='admissionregistration', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/main.libsonnet new file mode 100644 index 00000000000..c98e455e884 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/main.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + mutatingWebhook: (import 'mutatingWebhook.libsonnet'), + mutatingWebhookConfiguration: (import 'mutatingWebhookConfiguration.libsonnet'), + ruleWithOperations: (import 'ruleWithOperations.libsonnet'), + serviceReference: (import 'serviceReference.libsonnet'), + validatingWebhook: (import 'validatingWebhook.libsonnet'), + validatingWebhookConfiguration: (import 'validatingWebhookConfiguration.libsonnet'), + webhookClientConfig: (import 'webhookClientConfig.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhook.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhook.libsonnet new file mode 100644 index 00000000000..37e4067085b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhook.libsonnet @@ -0,0 +1,66 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='mutatingWebhook', url='', help='MutatingWebhook describes an admission webhook and the resources and operations it applies to.'), + '#clientConfig':: d.obj(help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + clientConfig: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { clientConfig+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { clientConfig+: { service+: { namespace: namespace } } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { clientConfig+: { service+: { path: path } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { clientConfig+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { clientConfig+: { caBundle: caBundle } }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { clientConfig+: { url: url } }, + }, + '#namespaceSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + namespaceSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#objectSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + objectSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { objectSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { objectSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { objectSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { objectSelector+: { matchLabels+: matchLabels } }, + }, + '#withAdmissionReviewVersions':: d.fn(help='AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy.', args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersions(admissionReviewVersions): { admissionReviewVersions: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withAdmissionReviewVersionsMixin':: d.fn(help='AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersionsMixin(admissionReviewVersions): { admissionReviewVersions+: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withFailurePolicy':: d.fn(help='FailurePolicy defines how unrecognized errors from the admission endpoint are handled - allowed values are Ignore or Fail. Defaults to Fail.', args=[d.arg(name='failurePolicy', type=d.T.string)]), + withFailurePolicy(failurePolicy): { failurePolicy: failurePolicy }, + '#withMatchPolicy':: d.fn(help='matchPolicy defines how the "rules" list is used to match incoming requests. Allowed values are "Exact" or "Equivalent".\n\n- Exact: match a request only if it exactly matches a specified rule. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, but "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would not be sent to the webhook.\n\n- Equivalent: match a request if modifies a resource listed in rules, even via another API group or version. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, and "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would be converted to apps/v1 and sent to the webhook.\n\nDefaults to "Equivalent"', args=[d.arg(name='matchPolicy', type=d.T.string)]), + withMatchPolicy(matchPolicy): { matchPolicy: matchPolicy }, + '#withName':: d.fn(help='The name of the admission webhook. Name should be fully qualified, e.g., imagepolicy.kubernetes.io, where "imagepolicy" is the name of the webhook, and kubernetes.io is the name of the organization. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReinvocationPolicy':: d.fn(help='reinvocationPolicy indicates whether this webhook should be called multiple times as part of a single admission evaluation. Allowed values are "Never" and "IfNeeded".\n\nNever: the webhook will not be called more than once in a single admission evaluation.\n\nIfNeeded: the webhook will be called at least one additional time as part of the admission evaluation if the object being admitted is modified by other admission plugins after the initial webhook call. Webhooks that specify this option *must* be idempotent, able to process objects they previously admitted. Note: * the number of additional invocations is not guaranteed to be exactly one. * if additional invocations result in further modifications to the object, webhooks are not guaranteed to be invoked again. * webhooks that use this option may be reordered to minimize the number of additional invocations. * to validate an object after all mutations are guaranteed complete, use a validating admission webhook instead.\n\nDefaults to "Never".', args=[d.arg(name='reinvocationPolicy', type=d.T.string)]), + withReinvocationPolicy(reinvocationPolicy): { reinvocationPolicy: reinvocationPolicy }, + '#withRules':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withSideEffects':: d.fn(help='SideEffects states whether this webhook has side effects. Acceptable values are: None, NoneOnDryRun (webhooks created via v1beta1 may also specify Some or Unknown). Webhooks with side effects MUST implement a reconciliation system, since a request may be rejected by a future step in the admission change and the side effects therefore need to be undone. Requests with the dryRun attribute will be auto-rejected if they match a webhook with sideEffects == Unknown or Some.', args=[d.arg(name='sideEffects', type=d.T.string)]), + withSideEffects(sideEffects): { sideEffects: sideEffects }, + '#withTimeoutSeconds':: d.fn(help='TimeoutSeconds specifies the timeout for this webhook. After the timeout passes, the webhook call will be ignored or the API call will fail based on the failure policy. The timeout value must be between 1 and 30 seconds. Default to 10 seconds.', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhookConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhookConfiguration.libsonnet new file mode 100644 index 00000000000..b0cdb4a8190 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/mutatingWebhookConfiguration.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='mutatingWebhookConfiguration', url='', help='MutatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and may change the object.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Mutatingwebhookconfiguration', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'admissionregistration.k8s.io/v1', + kind: 'MutatingWebhookConfiguration', + } + self.metadata.withName(name=name), + '#withWebhooks':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooks(webhooks): { webhooks: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#withWebhooksMixin':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooksMixin(webhooks): { webhooks+: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/ruleWithOperations.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/ruleWithOperations.libsonnet new file mode 100644 index 00000000000..a1661438992 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/ruleWithOperations.libsonnet @@ -0,0 +1,24 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ruleWithOperations', url='', help='RuleWithOperations is a tuple of Operations and Resources. It is recommended to make sure that all the tuple expansions are valid.'), + '#withApiGroups':: d.fn(help="APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help="APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiVersions':: d.fn(help="APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='apiVersions', type=d.T.array)]), + withApiVersions(apiVersions): { apiVersions: if std.isArray(v=apiVersions) then apiVersions else [apiVersions] }, + '#withApiVersionsMixin':: d.fn(help="APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='apiVersions', type=d.T.array)]), + withApiVersionsMixin(apiVersions): { apiVersions+: if std.isArray(v=apiVersions) then apiVersions else [apiVersions] }, + '#withOperations':: d.fn(help="Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='operations', type=d.T.array)]), + withOperations(operations): { operations: if std.isArray(v=operations) then operations else [operations] }, + '#withOperationsMixin':: d.fn(help="Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='operations', type=d.T.array)]), + withOperationsMixin(operations): { operations+: if std.isArray(v=operations) then operations else [operations] }, + '#withResources':: d.fn(help="Resources is a list of resources this rule applies to.\n\nFor example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources.\n\nIf wildcard is present, the validation rule will ensure resources do not overlap with each other.\n\nDepending on the enclosing object, subresources might not be allowed. Required.", args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help="Resources is a list of resources this rule applies to.\n\nFor example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources.\n\nIf wildcard is present, the validation rule will ensure resources do not overlap with each other.\n\nDepending on the enclosing object, subresources might not be allowed. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withScope':: d.fn(help='scope specifies the scope of this rule. Valid values are "Cluster", "Namespaced", and "*" "Cluster" means that only cluster-scoped resources will match this rule. Namespace API objects are cluster-scoped. "Namespaced" means that only namespaced resources will match this rule. "*" means that there are no scope restrictions. Subresources match the scope of their parent resource. Default is "*".', args=[d.arg(name='scope', type=d.T.string)]), + withScope(scope): { scope: scope }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/serviceReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/serviceReference.libsonnet new file mode 100644 index 00000000000..f210ec6515f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/serviceReference.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceReference', url='', help='ServiceReference holds a reference to Service.legacy.k8s.io'), + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhook.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhook.libsonnet new file mode 100644 index 00000000000..2e730a4ac95 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhook.libsonnet @@ -0,0 +1,64 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='validatingWebhook', url='', help='ValidatingWebhook describes an admission webhook and the resources and operations it applies to.'), + '#clientConfig':: d.obj(help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + clientConfig: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { clientConfig+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { clientConfig+: { service+: { namespace: namespace } } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { clientConfig+: { service+: { path: path } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { clientConfig+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { clientConfig+: { caBundle: caBundle } }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { clientConfig+: { url: url } }, + }, + '#namespaceSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + namespaceSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#objectSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + objectSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { objectSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { objectSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { objectSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { objectSelector+: { matchLabels+: matchLabels } }, + }, + '#withAdmissionReviewVersions':: d.fn(help='AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy.', args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersions(admissionReviewVersions): { admissionReviewVersions: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withAdmissionReviewVersionsMixin':: d.fn(help='AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersionsMixin(admissionReviewVersions): { admissionReviewVersions+: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withFailurePolicy':: d.fn(help='FailurePolicy defines how unrecognized errors from the admission endpoint are handled - allowed values are Ignore or Fail. Defaults to Fail.', args=[d.arg(name='failurePolicy', type=d.T.string)]), + withFailurePolicy(failurePolicy): { failurePolicy: failurePolicy }, + '#withMatchPolicy':: d.fn(help='matchPolicy defines how the "rules" list is used to match incoming requests. Allowed values are "Exact" or "Equivalent".\n\n- Exact: match a request only if it exactly matches a specified rule. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, but "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would not be sent to the webhook.\n\n- Equivalent: match a request if modifies a resource listed in rules, even via another API group or version. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, and "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would be converted to apps/v1 and sent to the webhook.\n\nDefaults to "Equivalent"', args=[d.arg(name='matchPolicy', type=d.T.string)]), + withMatchPolicy(matchPolicy): { matchPolicy: matchPolicy }, + '#withName':: d.fn(help='The name of the admission webhook. Name should be fully qualified, e.g., imagepolicy.kubernetes.io, where "imagepolicy" is the name of the webhook, and kubernetes.io is the name of the organization. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withRules':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withSideEffects':: d.fn(help='SideEffects states whether this webhook has side effects. Acceptable values are: None, NoneOnDryRun (webhooks created via v1beta1 may also specify Some or Unknown). Webhooks with side effects MUST implement a reconciliation system, since a request may be rejected by a future step in the admission change and the side effects therefore need to be undone. Requests with the dryRun attribute will be auto-rejected if they match a webhook with sideEffects == Unknown or Some.', args=[d.arg(name='sideEffects', type=d.T.string)]), + withSideEffects(sideEffects): { sideEffects: sideEffects }, + '#withTimeoutSeconds':: d.fn(help='TimeoutSeconds specifies the timeout for this webhook. After the timeout passes, the webhook call will be ignored or the API call will fail based on the failure policy. The timeout value must be between 1 and 30 seconds. Default to 10 seconds.', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhookConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhookConfiguration.libsonnet new file mode 100644 index 00000000000..fe9ea39824a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/validatingWebhookConfiguration.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='validatingWebhookConfiguration', url='', help='ValidatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and object without changing it.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Validatingwebhookconfiguration', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'admissionregistration.k8s.io/v1', + kind: 'ValidatingWebhookConfiguration', + } + self.metadata.withName(name=name), + '#withWebhooks':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooks(webhooks): { webhooks: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#withWebhooksMixin':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooksMixin(webhooks): { webhooks+: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/webhookClientConfig.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/webhookClientConfig.libsonnet new file mode 100644 index 00000000000..d0875aaec4e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1/webhookClientConfig.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='webhookClientConfig', url='', help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { name: name } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { service+: { namespace: namespace } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { service+: { path: path } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { service+: { port: port } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { caBundle: caBundle }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { url: url }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/main.libsonnet new file mode 100644 index 00000000000..8802f84a51d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/main.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + mutatingWebhook: (import 'mutatingWebhook.libsonnet'), + mutatingWebhookConfiguration: (import 'mutatingWebhookConfiguration.libsonnet'), + ruleWithOperations: (import 'ruleWithOperations.libsonnet'), + serviceReference: (import 'serviceReference.libsonnet'), + validatingWebhook: (import 'validatingWebhook.libsonnet'), + validatingWebhookConfiguration: (import 'validatingWebhookConfiguration.libsonnet'), + webhookClientConfig: (import 'webhookClientConfig.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhook.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhook.libsonnet new file mode 100644 index 00000000000..bc714e12e14 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhook.libsonnet @@ -0,0 +1,66 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='mutatingWebhook', url='', help='MutatingWebhook describes an admission webhook and the resources and operations it applies to.'), + '#clientConfig':: d.obj(help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + clientConfig: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { clientConfig+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { clientConfig+: { service+: { namespace: namespace } } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { clientConfig+: { service+: { path: path } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { clientConfig+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { clientConfig+: { caBundle: caBundle } }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { clientConfig+: { url: url } }, + }, + '#namespaceSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + namespaceSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#objectSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + objectSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { objectSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { objectSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { objectSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { objectSelector+: { matchLabels+: matchLabels } }, + }, + '#withAdmissionReviewVersions':: d.fn(help="AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`.", args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersions(admissionReviewVersions): { admissionReviewVersions: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withAdmissionReviewVersionsMixin':: d.fn(help="AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersionsMixin(admissionReviewVersions): { admissionReviewVersions+: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withFailurePolicy':: d.fn(help='FailurePolicy defines how unrecognized errors from the admission endpoint are handled - allowed values are Ignore or Fail. Defaults to Ignore.', args=[d.arg(name='failurePolicy', type=d.T.string)]), + withFailurePolicy(failurePolicy): { failurePolicy: failurePolicy }, + '#withMatchPolicy':: d.fn(help='matchPolicy defines how the "rules" list is used to match incoming requests. Allowed values are "Exact" or "Equivalent".\n\n- Exact: match a request only if it exactly matches a specified rule. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, but "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would not be sent to the webhook.\n\n- Equivalent: match a request if modifies a resource listed in rules, even via another API group or version. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, and "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would be converted to apps/v1 and sent to the webhook.\n\nDefaults to "Exact"', args=[d.arg(name='matchPolicy', type=d.T.string)]), + withMatchPolicy(matchPolicy): { matchPolicy: matchPolicy }, + '#withName':: d.fn(help='The name of the admission webhook. Name should be fully qualified, e.g., imagepolicy.kubernetes.io, where "imagepolicy" is the name of the webhook, and kubernetes.io is the name of the organization. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReinvocationPolicy':: d.fn(help='reinvocationPolicy indicates whether this webhook should be called multiple times as part of a single admission evaluation. Allowed values are "Never" and "IfNeeded".\n\nNever: the webhook will not be called more than once in a single admission evaluation.\n\nIfNeeded: the webhook will be called at least one additional time as part of the admission evaluation if the object being admitted is modified by other admission plugins after the initial webhook call. Webhooks that specify this option *must* be idempotent, able to process objects they previously admitted. Note: * the number of additional invocations is not guaranteed to be exactly one. * if additional invocations result in further modifications to the object, webhooks are not guaranteed to be invoked again. * webhooks that use this option may be reordered to minimize the number of additional invocations. * to validate an object after all mutations are guaranteed complete, use a validating admission webhook instead.\n\nDefaults to "Never".', args=[d.arg(name='reinvocationPolicy', type=d.T.string)]), + withReinvocationPolicy(reinvocationPolicy): { reinvocationPolicy: reinvocationPolicy }, + '#withRules':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withSideEffects':: d.fn(help='SideEffects states whether this webhook has side effects. Acceptable values are: Unknown, None, Some, NoneOnDryRun Webhooks with side effects MUST implement a reconciliation system, since a request may be rejected by a future step in the admission change and the side effects therefore need to be undone. Requests with the dryRun attribute will be auto-rejected if they match a webhook with sideEffects == Unknown or Some. Defaults to Unknown.', args=[d.arg(name='sideEffects', type=d.T.string)]), + withSideEffects(sideEffects): { sideEffects: sideEffects }, + '#withTimeoutSeconds':: d.fn(help='TimeoutSeconds specifies the timeout for this webhook. After the timeout passes, the webhook call will be ignored or the API call will fail based on the failure policy. The timeout value must be between 1 and 30 seconds. Default to 30 seconds.', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhookConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhookConfiguration.libsonnet new file mode 100644 index 00000000000..f04f9486c1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/mutatingWebhookConfiguration.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='mutatingWebhookConfiguration', url='', help='MutatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and may change the object. Deprecated in v1.16, planned for removal in v1.19. Use admissionregistration.k8s.io/v1 MutatingWebhookConfiguration instead.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Mutatingwebhookconfiguration', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'admissionregistration.k8s.io/v1beta1', + kind: 'MutatingWebhookConfiguration', + } + self.metadata.withName(name=name), + '#withWebhooks':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooks(webhooks): { webhooks: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#withWebhooksMixin':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooksMixin(webhooks): { webhooks+: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/ruleWithOperations.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/ruleWithOperations.libsonnet new file mode 100644 index 00000000000..a1661438992 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/ruleWithOperations.libsonnet @@ -0,0 +1,24 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ruleWithOperations', url='', help='RuleWithOperations is a tuple of Operations and Resources. It is recommended to make sure that all the tuple expansions are valid.'), + '#withApiGroups':: d.fn(help="APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help="APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiVersions':: d.fn(help="APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='apiVersions', type=d.T.array)]), + withApiVersions(apiVersions): { apiVersions: if std.isArray(v=apiVersions) then apiVersions else [apiVersions] }, + '#withApiVersionsMixin':: d.fn(help="APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='apiVersions', type=d.T.array)]), + withApiVersionsMixin(apiVersions): { apiVersions+: if std.isArray(v=apiVersions) then apiVersions else [apiVersions] }, + '#withOperations':: d.fn(help="Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.", args=[d.arg(name='operations', type=d.T.array)]), + withOperations(operations): { operations: if std.isArray(v=operations) then operations else [operations] }, + '#withOperationsMixin':: d.fn(help="Operations is the operations the admission hook cares about - CREATE, UPDATE, DELETE, CONNECT or * for all of those operations and any future admission operations that are added. If '*' is present, the length of the slice must be one. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='operations', type=d.T.array)]), + withOperationsMixin(operations): { operations+: if std.isArray(v=operations) then operations else [operations] }, + '#withResources':: d.fn(help="Resources is a list of resources this rule applies to.\n\nFor example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources.\n\nIf wildcard is present, the validation rule will ensure resources do not overlap with each other.\n\nDepending on the enclosing object, subresources might not be allowed. Required.", args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help="Resources is a list of resources this rule applies to.\n\nFor example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources.\n\nIf wildcard is present, the validation rule will ensure resources do not overlap with each other.\n\nDepending on the enclosing object, subresources might not be allowed. Required.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withScope':: d.fn(help='scope specifies the scope of this rule. Valid values are "Cluster", "Namespaced", and "*" "Cluster" means that only cluster-scoped resources will match this rule. Namespace API objects are cluster-scoped. "Namespaced" means that only namespaced resources will match this rule. "*" means that there are no scope restrictions. Subresources match the scope of their parent resource. Default is "*".', args=[d.arg(name='scope', type=d.T.string)]), + withScope(scope): { scope: scope }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/serviceReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/serviceReference.libsonnet new file mode 100644 index 00000000000..f210ec6515f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/serviceReference.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceReference', url='', help='ServiceReference holds a reference to Service.legacy.k8s.io'), + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhook.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhook.libsonnet new file mode 100644 index 00000000000..1d7ea945447 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhook.libsonnet @@ -0,0 +1,64 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='validatingWebhook', url='', help='ValidatingWebhook describes an admission webhook and the resources and operations it applies to.'), + '#clientConfig':: d.obj(help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + clientConfig: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { clientConfig+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { clientConfig+: { service+: { namespace: namespace } } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { clientConfig+: { service+: { path: path } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { clientConfig+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { clientConfig+: { caBundle: caBundle } }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { clientConfig+: { url: url } }, + }, + '#namespaceSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + namespaceSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#objectSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + objectSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { objectSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { objectSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { objectSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { objectSelector+: { matchLabels+: matchLabels } }, + }, + '#withAdmissionReviewVersions':: d.fn(help="AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`.", args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersions(admissionReviewVersions): { admissionReviewVersions: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withAdmissionReviewVersionsMixin':: d.fn(help="AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='admissionReviewVersions', type=d.T.array)]), + withAdmissionReviewVersionsMixin(admissionReviewVersions): { admissionReviewVersions+: if std.isArray(v=admissionReviewVersions) then admissionReviewVersions else [admissionReviewVersions] }, + '#withFailurePolicy':: d.fn(help='FailurePolicy defines how unrecognized errors from the admission endpoint are handled - allowed values are Ignore or Fail. Defaults to Ignore.', args=[d.arg(name='failurePolicy', type=d.T.string)]), + withFailurePolicy(failurePolicy): { failurePolicy: failurePolicy }, + '#withMatchPolicy':: d.fn(help='matchPolicy defines how the "rules" list is used to match incoming requests. Allowed values are "Exact" or "Equivalent".\n\n- Exact: match a request only if it exactly matches a specified rule. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, but "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would not be sent to the webhook.\n\n- Equivalent: match a request if modifies a resource listed in rules, even via another API group or version. For example, if deployments can be modified via apps/v1, apps/v1beta1, and extensions/v1beta1, and "rules" only included `apiGroups:["apps"], apiVersions:["v1"], resources: ["deployments"]`, a request to apps/v1beta1 or extensions/v1beta1 would be converted to apps/v1 and sent to the webhook.\n\nDefaults to "Exact"', args=[d.arg(name='matchPolicy', type=d.T.string)]), + withMatchPolicy(matchPolicy): { matchPolicy: matchPolicy }, + '#withName':: d.fn(help='The name of the admission webhook. Name should be fully qualified, e.g., imagepolicy.kubernetes.io, where "imagepolicy" is the name of the webhook, and kubernetes.io is the name of the organization. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withRules':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withSideEffects':: d.fn(help='SideEffects states whether this webhook has side effects. Acceptable values are: Unknown, None, Some, NoneOnDryRun Webhooks with side effects MUST implement a reconciliation system, since a request may be rejected by a future step in the admission change and the side effects therefore need to be undone. Requests with the dryRun attribute will be auto-rejected if they match a webhook with sideEffects == Unknown or Some. Defaults to Unknown.', args=[d.arg(name='sideEffects', type=d.T.string)]), + withSideEffects(sideEffects): { sideEffects: sideEffects }, + '#withTimeoutSeconds':: d.fn(help='TimeoutSeconds specifies the timeout for this webhook. After the timeout passes, the webhook call will be ignored or the API call will fail based on the failure policy. The timeout value must be between 1 and 30 seconds. Default to 30 seconds.', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhookConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhookConfiguration.libsonnet new file mode 100644 index 00000000000..728bebc9876 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/validatingWebhookConfiguration.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='validatingWebhookConfiguration', url='', help='ValidatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and object without changing it. Deprecated in v1.16, planned for removal in v1.19. Use admissionregistration.k8s.io/v1 ValidatingWebhookConfiguration instead.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Validatingwebhookconfiguration', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'admissionregistration.k8s.io/v1beta1', + kind: 'ValidatingWebhookConfiguration', + } + self.metadata.withName(name=name), + '#withWebhooks':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooks(webhooks): { webhooks: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#withWebhooksMixin':: d.fn(help='Webhooks is a list of webhooks and the affected resources and operations.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='webhooks', type=d.T.array)]), + withWebhooksMixin(webhooks): { webhooks+: if std.isArray(v=webhooks) then webhooks else [webhooks] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/webhookClientConfig.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/webhookClientConfig.libsonnet new file mode 100644 index 00000000000..d0875aaec4e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/admissionregistration/v1beta1/webhookClientConfig.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='webhookClientConfig', url='', help='WebhookClientConfig contains the information to make a TLS connection with the webhook'), + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='`name` is the name of the service. Required', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { name: name } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of the service. Required', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { service+: { namespace: namespace } }, + '#withPath':: d.fn(help='`path` is an optional URL path which will be sent in any request to this service.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { service+: { path: path } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { service+: { port: port } }, + }, + '#withCaBundle':: d.fn(help="`caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { caBundle: caBundle }, + '#withUrl':: d.fn(help='`url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified.\n\nThe `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address.\n\nPlease note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster.\n\nThe scheme must be "https"; the URL must begin with "https://".\n\nA path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier.\n\nAttempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either.', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { url: url }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/main.libsonnet new file mode 100644 index 00000000000..f4839e1e05b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiregistration', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiService.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiService.libsonnet new file mode 100644 index 00000000000..e92c601ab12 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiService.libsonnet @@ -0,0 +1,80 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiService', url='', help='APIService represents a server for a particular GroupVersion. Name must be "version.group".'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Apiservice', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apiregistration.k8s.io/v1', + kind: 'ApiService', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification.'), + spec: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { service+: { namespace: namespace } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { spec+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { spec+: { caBundle: caBundle } }, + '#withGroup':: d.fn(help='Group is the API group name this server hosts', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { group: group } }, + '#withGroupPriorityMinimum':: d.fn(help="GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s", args=[d.arg(name='groupPriorityMinimum', type=d.T.integer)]), + withGroupPriorityMinimum(groupPriorityMinimum): { spec+: { groupPriorityMinimum: groupPriorityMinimum } }, + '#withInsecureSkipTLSVerify':: d.fn(help='InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead.', args=[d.arg(name='insecureSkipTLSVerify', type=d.T.boolean)]), + withInsecureSkipTLSVerify(insecureSkipTLSVerify): { spec+: { insecureSkipTLSVerify: insecureSkipTLSVerify } }, + '#withVersion':: d.fn(help='Version is the API version this server hosts. For example, "v1"', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { version: version } }, + '#withVersionPriority':: d.fn(help="VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is 'kube-like', it will sort above non 'kube-like' version strings, which are ordered lexicographically. 'Kube-like' versions start with a 'v', then are followed by a number (the major version), then optionally the string 'alpha' or 'beta' and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10.", args=[d.arg(name='versionPriority', type=d.T.integer)]), + withVersionPriority(versionPriority): { spec+: { versionPriority: versionPriority } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceCondition.libsonnet new file mode 100644 index 00000000000..916decaecd2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceCondition', url='', help='APIServiceCondition describes the state of an APIService at a particular point'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="Unique, one-word, CamelCase reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type is the type of the condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceSpec.libsonnet new file mode 100644 index 00000000000..dc2df9fd1f9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceSpec.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceSpec', url='', help='APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification.'), + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { service+: { namespace: namespace } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { service+: { port: port } }, + }, + '#withCaBundle':: d.fn(help="CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { caBundle: caBundle }, + '#withGroup':: d.fn(help='Group is the API group name this server hosts', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withGroupPriorityMinimum':: d.fn(help="GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s", args=[d.arg(name='groupPriorityMinimum', type=d.T.integer)]), + withGroupPriorityMinimum(groupPriorityMinimum): { groupPriorityMinimum: groupPriorityMinimum }, + '#withInsecureSkipTLSVerify':: d.fn(help='InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead.', args=[d.arg(name='insecureSkipTLSVerify', type=d.T.boolean)]), + withInsecureSkipTLSVerify(insecureSkipTLSVerify): { insecureSkipTLSVerify: insecureSkipTLSVerify }, + '#withVersion':: d.fn(help='Version is the API version this server hosts. For example, "v1"', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { version: version }, + '#withVersionPriority':: d.fn(help="VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is 'kube-like', it will sort above non 'kube-like' version strings, which are ordered lexicographically. 'Kube-like' versions start with a 'v', then are followed by a number (the major version), then optionally the string 'alpha' or 'beta' and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10.", args=[d.arg(name='versionPriority', type=d.T.integer)]), + withVersionPriority(versionPriority): { versionPriority: versionPriority }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceStatus.libsonnet new file mode 100644 index 00000000000..00662ba98e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/apiServiceStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceStatus', url='', help='APIServiceStatus contains derived information about an API server'), + '#withConditions':: d.fn(help='Current service state of apiService.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='Current service state of apiService.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/main.libsonnet new file mode 100644 index 00000000000..a5c6e833dd2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/main.libsonnet @@ -0,0 +1,9 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + apiService: (import 'apiService.libsonnet'), + apiServiceCondition: (import 'apiServiceCondition.libsonnet'), + apiServiceSpec: (import 'apiServiceSpec.libsonnet'), + apiServiceStatus: (import 'apiServiceStatus.libsonnet'), + serviceReference: (import 'serviceReference.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/serviceReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/serviceReference.libsonnet new file mode 100644 index 00000000000..2a9557db86c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1/serviceReference.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceReference', url='', help='ServiceReference holds a reference to Service.legacy.k8s.io'), + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiService.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiService.libsonnet new file mode 100644 index 00000000000..bf89dc71d4b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiService.libsonnet @@ -0,0 +1,80 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiService', url='', help='APIService represents a server for a particular GroupVersion. Name must be "version.group".'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Apiservice', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apiregistration.k8s.io/v1beta1', + kind: 'ApiService', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification.'), + spec: { + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { service+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { service+: { namespace: namespace } } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { spec+: { service+: { port: port } } }, + }, + '#withCaBundle':: d.fn(help="CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { spec+: { caBundle: caBundle } }, + '#withGroup':: d.fn(help='Group is the API group name this server hosts', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { group: group } }, + '#withGroupPriorityMinimum':: d.fn(help="GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s", args=[d.arg(name='groupPriorityMinimum', type=d.T.integer)]), + withGroupPriorityMinimum(groupPriorityMinimum): { spec+: { groupPriorityMinimum: groupPriorityMinimum } }, + '#withInsecureSkipTLSVerify':: d.fn(help='InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead.', args=[d.arg(name='insecureSkipTLSVerify', type=d.T.boolean)]), + withInsecureSkipTLSVerify(insecureSkipTLSVerify): { spec+: { insecureSkipTLSVerify: insecureSkipTLSVerify } }, + '#withVersion':: d.fn(help='Version is the API version this server hosts. For example, "v1"', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { version: version } }, + '#withVersionPriority':: d.fn(help="VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is 'kube-like', it will sort above non 'kube-like' version strings, which are ordered lexicographically. 'Kube-like' versions start with a 'v', then are followed by a number (the major version), then optionally the string 'alpha' or 'beta' and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10.", args=[d.arg(name='versionPriority', type=d.T.integer)]), + withVersionPriority(versionPriority): { spec+: { versionPriority: versionPriority } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceCondition.libsonnet new file mode 100644 index 00000000000..916decaecd2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceCondition', url='', help='APIServiceCondition describes the state of an APIService at a particular point'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="Unique, one-word, CamelCase reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type is the type of the condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceSpec.libsonnet new file mode 100644 index 00000000000..dc2df9fd1f9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceSpec.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceSpec', url='', help='APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification.'), + '#service':: d.obj(help='ServiceReference holds a reference to Service.legacy.k8s.io'), + service: { + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { service+: { namespace: namespace } }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { service+: { port: port } }, + }, + '#withCaBundle':: d.fn(help="CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used.", args=[d.arg(name='caBundle', type=d.T.string)]), + withCaBundle(caBundle): { caBundle: caBundle }, + '#withGroup':: d.fn(help='Group is the API group name this server hosts', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withGroupPriorityMinimum':: d.fn(help="GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s", args=[d.arg(name='groupPriorityMinimum', type=d.T.integer)]), + withGroupPriorityMinimum(groupPriorityMinimum): { groupPriorityMinimum: groupPriorityMinimum }, + '#withInsecureSkipTLSVerify':: d.fn(help='InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead.', args=[d.arg(name='insecureSkipTLSVerify', type=d.T.boolean)]), + withInsecureSkipTLSVerify(insecureSkipTLSVerify): { insecureSkipTLSVerify: insecureSkipTLSVerify }, + '#withVersion':: d.fn(help='Version is the API version this server hosts. For example, "v1"', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { version: version }, + '#withVersionPriority':: d.fn(help="VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is 'kube-like', it will sort above non 'kube-like' version strings, which are ordered lexicographically. 'Kube-like' versions start with a 'v', then are followed by a number (the major version), then optionally the string 'alpha' or 'beta' and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10.", args=[d.arg(name='versionPriority', type=d.T.integer)]), + withVersionPriority(versionPriority): { versionPriority: versionPriority }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceStatus.libsonnet new file mode 100644 index 00000000000..00662ba98e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/apiServiceStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apiServiceStatus', url='', help='APIServiceStatus contains derived information about an API server'), + '#withConditions':: d.fn(help='Current service state of apiService.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='Current service state of apiService.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/main.libsonnet new file mode 100644 index 00000000000..2cdc33f58df --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/main.libsonnet @@ -0,0 +1,9 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + apiService: (import 'apiService.libsonnet'), + apiServiceCondition: (import 'apiServiceCondition.libsonnet'), + apiServiceSpec: (import 'apiServiceSpec.libsonnet'), + apiServiceStatus: (import 'apiServiceStatus.libsonnet'), + serviceReference: (import 'serviceReference.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/serviceReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/serviceReference.libsonnet new file mode 100644 index 00000000000..2a9557db86c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apiregistration/v1beta1/serviceReference.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceReference', url='', help='ServiceReference holds a reference to Service.legacy.k8s.io'), + '#withName':: d.fn(help='Name is the name of the service', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the service', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withPort':: d.fn(help='If specified, the port on the service that hosting webhook. Default to 443 for backward compatibility. `port` should be a valid port number (1-65535, inclusive).', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/main.libsonnet new file mode 100644 index 00000000000..36337c27d18 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='apps', url='', help=''), + v1: (import 'v1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/controllerRevision.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/controllerRevision.libsonnet new file mode 100644 index 00000000000..251deb72e60 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/controllerRevision.libsonnet @@ -0,0 +1,62 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='controllerRevision', url='', help='ControllerRevision implements an immutable snapshot of state data. Clients are responsible for serializing and deserializing the objects that contain their internal state. Once a ControllerRevision has been successfully created, it can not be updated. The API Server will fail validation of all requests that attempt to mutate the Data field. ControllerRevisions may, however, be deleted. Note that, due to its use by both the DaemonSet and StatefulSet controllers for update and rollback, this object is beta. However, it may be subject to name and representation changes in future releases, and clients should not depend on its stability. It is primarily for internal use by controllers.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Controllerrevision', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apps/v1', + kind: 'ControllerRevision', + } + self.metadata.withName(name=name), + '#withData':: d.fn(help="RawExtension is used to hold extensions in external versions.\n\nTo use this, make a field which has RawExtension as its type in your external, versioned struct, and Object in your internal struct. You also need to register your various plugin types.\n\n// Internal package: type MyAPIObject struct {\n\truntime.TypeMeta `json:',inline'`\n\tMyPlugin runtime.Object `json:'myPlugin'`\n} type PluginA struct {\n\tAOption string `json:'aOption'`\n}\n\n// External package: type MyAPIObject struct {\n\truntime.TypeMeta `json:',inline'`\n\tMyPlugin runtime.RawExtension `json:'myPlugin'`\n} type PluginA struct {\n\tAOption string `json:'aOption'`\n}\n\n// On the wire, the JSON will look something like this: {\n\t'kind':'MyAPIObject',\n\t'apiVersion':'v1',\n\t'myPlugin': {\n\t\t'kind':'PluginA',\n\t\t'aOption':'foo',\n\t},\n}\n\nSo what happens? Decode first uses json or yaml to unmarshal the serialized data into your external MyAPIObject. That causes the raw JSON to be stored, but not unpacked. The next step is to copy (using pkg/conversion) into the internal struct. The runtime package's DefaultScheme has conversion functions installed which will unpack the JSON stored in RawExtension, turning it into the correct object type, and storing it in the Object. (TODO: In the case where the object is of an unknown type, a runtime.Unknown object will be created and stored.)", args=[d.arg(name='data', type=d.T.object)]), + withData(data): { data: data }, + '#withDataMixin':: d.fn(help="RawExtension is used to hold extensions in external versions.\n\nTo use this, make a field which has RawExtension as its type in your external, versioned struct, and Object in your internal struct. You also need to register your various plugin types.\n\n// Internal package: type MyAPIObject struct {\n\truntime.TypeMeta `json:',inline'`\n\tMyPlugin runtime.Object `json:'myPlugin'`\n} type PluginA struct {\n\tAOption string `json:'aOption'`\n}\n\n// External package: type MyAPIObject struct {\n\truntime.TypeMeta `json:',inline'`\n\tMyPlugin runtime.RawExtension `json:'myPlugin'`\n} type PluginA struct {\n\tAOption string `json:'aOption'`\n}\n\n// On the wire, the JSON will look something like this: {\n\t'kind':'MyAPIObject',\n\t'apiVersion':'v1',\n\t'myPlugin': {\n\t\t'kind':'PluginA',\n\t\t'aOption':'foo',\n\t},\n}\n\nSo what happens? Decode first uses json or yaml to unmarshal the serialized data into your external MyAPIObject. That causes the raw JSON to be stored, but not unpacked. The next step is to copy (using pkg/conversion) into the internal struct. The runtime package's DefaultScheme has conversion functions installed which will unpack the JSON stored in RawExtension, turning it into the correct object type, and storing it in the Object. (TODO: In the case where the object is of an unknown type, a runtime.Unknown object will be created and stored.)\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='data', type=d.T.object)]), + withDataMixin(data): { data+: data }, + '#withRevision':: d.fn(help='Revision indicates the revision of the state represented by Data.', args=[d.arg(name='revision', type=d.T.integer)]), + withRevision(revision): { revision: revision }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSet.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSet.libsonnet new file mode 100644 index 00000000000..9a33c500052 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSet.libsonnet @@ -0,0 +1,323 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonSet', url='', help='DaemonSet represents the configuration of a daemon set.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Daemonset', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apps/v1', + kind: 'DaemonSet', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='DaemonSetSpec is the specification of a daemon set.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#updateStrategy':: d.obj(help='DaemonSetUpdateStrategy is a struct used to control the update strategy for a DaemonSet.'), + updateStrategy: { + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of daemon set rolling update.'), + rollingUpdate: { + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { spec+: { updateStrategy+: { rollingUpdate+: { maxUnavailable: maxUnavailable } } } }, + }, + '#withType':: d.fn(help='Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { updateStrategy+: { type: type } } }, + }, + '#withMinReadySeconds':: d.fn(help='The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready).', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withRevisionHistoryLimit':: d.fn(help='The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10.', args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { spec+: { revisionHistoryLimit: revisionHistoryLimit } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetCondition.libsonnet new file mode 100644 index 00000000000..cf9d64b2669 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonSetCondition', url='', help='DaemonSetCondition describes the state of a DaemonSet at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='A human readable message indicating details about the transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="The reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of DaemonSet condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetSpec.libsonnet new file mode 100644 index 00000000000..d945aa75833 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetSpec.libsonnet @@ -0,0 +1,270 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonSetSpec', url='', help='DaemonSetSpec is the specification of a daemon set.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#updateStrategy':: d.obj(help='DaemonSetUpdateStrategy is a struct used to control the update strategy for a DaemonSet.'), + updateStrategy: { + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of daemon set rolling update.'), + rollingUpdate: { + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { updateStrategy+: { rollingUpdate+: { maxUnavailable: maxUnavailable } } }, + }, + '#withType':: d.fn(help='Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { updateStrategy+: { type: type } }, + }, + '#withMinReadySeconds':: d.fn(help='The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready).', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { minReadySeconds: minReadySeconds }, + '#withRevisionHistoryLimit':: d.fn(help='The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10.', args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { revisionHistoryLimit: revisionHistoryLimit }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetStatus.libsonnet new file mode 100644 index 00000000000..d0ee7972dad --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetStatus.libsonnet @@ -0,0 +1,28 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonSetStatus', url='', help='DaemonSetStatus represents the current status of a daemon set.'), + '#withCollisionCount':: d.fn(help='Count of hash collisions for the DaemonSet. The DaemonSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision.', args=[d.arg(name='collisionCount', type=d.T.integer)]), + withCollisionCount(collisionCount): { collisionCount: collisionCount }, + '#withConditions':: d.fn(help="Represents the latest available observations of a DaemonSet's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a DaemonSet's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withCurrentNumberScheduled':: d.fn(help='The number of nodes that are running at least 1 daemon pod and are supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/', args=[d.arg(name='currentNumberScheduled', type=d.T.integer)]), + withCurrentNumberScheduled(currentNumberScheduled): { currentNumberScheduled: currentNumberScheduled }, + '#withDesiredNumberScheduled':: d.fn(help='The total number of nodes that should be running the daemon pod (including nodes correctly running the daemon pod). More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/', args=[d.arg(name='desiredNumberScheduled', type=d.T.integer)]), + withDesiredNumberScheduled(desiredNumberScheduled): { desiredNumberScheduled: desiredNumberScheduled }, + '#withNumberAvailable':: d.fn(help='The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and available (ready for at least spec.minReadySeconds)', args=[d.arg(name='numberAvailable', type=d.T.integer)]), + withNumberAvailable(numberAvailable): { numberAvailable: numberAvailable }, + '#withNumberMisscheduled':: d.fn(help='The number of nodes that are running the daemon pod, but are not supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/', args=[d.arg(name='numberMisscheduled', type=d.T.integer)]), + withNumberMisscheduled(numberMisscheduled): { numberMisscheduled: numberMisscheduled }, + '#withNumberReady':: d.fn(help='The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and ready.', args=[d.arg(name='numberReady', type=d.T.integer)]), + withNumberReady(numberReady): { numberReady: numberReady }, + '#withNumberUnavailable':: d.fn(help='The number of nodes that should be running the daemon pod and have none of the daemon pod running and available (ready for at least spec.minReadySeconds)', args=[d.arg(name='numberUnavailable', type=d.T.integer)]), + withNumberUnavailable(numberUnavailable): { numberUnavailable: numberUnavailable }, + '#withObservedGeneration':: d.fn(help='The most recent generation observed by the daemon set controller.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#withUpdatedNumberScheduled':: d.fn(help='The total number of nodes that are running updated daemon pod', args=[d.arg(name='updatedNumberScheduled', type=d.T.integer)]), + withUpdatedNumberScheduled(updatedNumberScheduled): { updatedNumberScheduled: updatedNumberScheduled }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetUpdateStrategy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetUpdateStrategy.libsonnet new file mode 100644 index 00000000000..8eedb26db82 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/daemonSetUpdateStrategy.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonSetUpdateStrategy', url='', help='DaemonSetUpdateStrategy is a struct used to control the update strategy for a DaemonSet.'), + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of daemon set rolling update.'), + rollingUpdate: { + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { rollingUpdate+: { maxUnavailable: maxUnavailable } }, + }, + '#withType':: d.fn(help='Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deployment.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deployment.libsonnet new file mode 100644 index 00000000000..2ebb44419a7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deployment.libsonnet @@ -0,0 +1,331 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='deployment', url='', help='Deployment enables declarative updates for Pods and ReplicaSets.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Deployment', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apps/v1', + kind: 'Deployment', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='DeploymentSpec is the specification of the desired behavior of the Deployment.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#strategy':: d.obj(help='DeploymentStrategy describes how to replace existing pods with new ones.'), + strategy: { + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of rolling update.'), + rollingUpdate: { + '#withMaxSurge':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxSurge', type=d.T.string)]), + withMaxSurge(maxSurge): { spec+: { strategy+: { rollingUpdate+: { maxSurge: maxSurge } } } }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { spec+: { strategy+: { rollingUpdate+: { maxUnavailable: maxUnavailable } } } }, + }, + '#withType':: d.fn(help='Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { strategy+: { type: type } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withPaused':: d.fn(help='Indicates that the deployment is paused.', args=[d.arg(name='paused', type=d.T.boolean)]), + withPaused(paused): { spec+: { paused: paused } }, + '#withProgressDeadlineSeconds':: d.fn(help='The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s.', args=[d.arg(name='progressDeadlineSeconds', type=d.T.integer)]), + withProgressDeadlineSeconds(progressDeadlineSeconds): { spec+: { progressDeadlineSeconds: progressDeadlineSeconds } }, + '#withReplicas':: d.fn(help='Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + '#withRevisionHistoryLimit':: d.fn(help='The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10.', args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { spec+: { revisionHistoryLimit: revisionHistoryLimit } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentCondition.libsonnet new file mode 100644 index 00000000000..7b0bb30f44f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='deploymentCondition', url='', help='DeploymentCondition describes the state of a deployment at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withLastUpdateTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastUpdateTime', type=d.T.string)]), + withLastUpdateTime(lastUpdateTime): { lastUpdateTime: lastUpdateTime }, + '#withMessage':: d.fn(help='A human readable message indicating details about the transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="The reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of deployment condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentSpec.libsonnet new file mode 100644 index 00000000000..029cfc5456b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentSpec.libsonnet @@ -0,0 +1,278 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='deploymentSpec', url='', help='DeploymentSpec is the specification of the desired behavior of the Deployment.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#strategy':: d.obj(help='DeploymentStrategy describes how to replace existing pods with new ones.'), + strategy: { + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of rolling update.'), + rollingUpdate: { + '#withMaxSurge':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxSurge', type=d.T.string)]), + withMaxSurge(maxSurge): { strategy+: { rollingUpdate+: { maxSurge: maxSurge } } }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { strategy+: { rollingUpdate+: { maxUnavailable: maxUnavailable } } }, + }, + '#withType':: d.fn(help='Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { strategy+: { type: type } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { minReadySeconds: minReadySeconds }, + '#withPaused':: d.fn(help='Indicates that the deployment is paused.', args=[d.arg(name='paused', type=d.T.boolean)]), + withPaused(paused): { paused: paused }, + '#withProgressDeadlineSeconds':: d.fn(help='The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s.', args=[d.arg(name='progressDeadlineSeconds', type=d.T.integer)]), + withProgressDeadlineSeconds(progressDeadlineSeconds): { progressDeadlineSeconds: progressDeadlineSeconds }, + '#withReplicas':: d.fn(help='Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withRevisionHistoryLimit':: d.fn(help='The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10.', args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { revisionHistoryLimit: revisionHistoryLimit }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStatus.libsonnet new file mode 100644 index 00000000000..30eed867943 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStatus.libsonnet @@ -0,0 +1,24 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='deploymentStatus', url='', help='DeploymentStatus is the most recently observed status of the Deployment.'), + '#withAvailableReplicas':: d.fn(help='Total number of available pods (ready for at least minReadySeconds) targeted by this deployment.', args=[d.arg(name='availableReplicas', type=d.T.integer)]), + withAvailableReplicas(availableReplicas): { availableReplicas: availableReplicas }, + '#withCollisionCount':: d.fn(help='Count of hash collisions for the Deployment. The Deployment controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ReplicaSet.', args=[d.arg(name='collisionCount', type=d.T.integer)]), + withCollisionCount(collisionCount): { collisionCount: collisionCount }, + '#withConditions':: d.fn(help="Represents the latest available observations of a deployment's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a deployment's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withObservedGeneration':: d.fn(help='The generation observed by the deployment controller.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#withReadyReplicas':: d.fn(help='Total number of ready pods targeted by this deployment.', args=[d.arg(name='readyReplicas', type=d.T.integer)]), + withReadyReplicas(readyReplicas): { readyReplicas: readyReplicas }, + '#withReplicas':: d.fn(help='Total number of non-terminated pods targeted by this deployment (their labels match the selector).', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withUnavailableReplicas':: d.fn(help='Total number of unavailable pods targeted by this deployment. This is the total number of pods that are still required for the deployment to have 100% available capacity. They may either be pods that are running but not yet available or pods that still have not been created.', args=[d.arg(name='unavailableReplicas', type=d.T.integer)]), + withUnavailableReplicas(unavailableReplicas): { unavailableReplicas: unavailableReplicas }, + '#withUpdatedReplicas':: d.fn(help='Total number of non-terminated pods targeted by this deployment that have the desired template spec.', args=[d.arg(name='updatedReplicas', type=d.T.integer)]), + withUpdatedReplicas(updatedReplicas): { updatedReplicas: updatedReplicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStrategy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStrategy.libsonnet new file mode 100644 index 00000000000..aebf980782d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/deploymentStrategy.libsonnet @@ -0,0 +1,15 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='deploymentStrategy', url='', help='DeploymentStrategy describes how to replace existing pods with new ones.'), + '#rollingUpdate':: d.obj(help='Spec to control the desired behavior of rolling update.'), + rollingUpdate: { + '#withMaxSurge':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxSurge', type=d.T.string)]), + withMaxSurge(maxSurge): { rollingUpdate+: { maxSurge: maxSurge } }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { rollingUpdate+: { maxUnavailable: maxUnavailable } }, + }, + '#withType':: d.fn(help='Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/main.libsonnet new file mode 100644 index 00000000000..beecb814b3b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/main.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + controllerRevision: (import 'controllerRevision.libsonnet'), + daemonSet: (import 'daemonSet.libsonnet'), + daemonSetCondition: (import 'daemonSetCondition.libsonnet'), + daemonSetSpec: (import 'daemonSetSpec.libsonnet'), + daemonSetStatus: (import 'daemonSetStatus.libsonnet'), + daemonSetUpdateStrategy: (import 'daemonSetUpdateStrategy.libsonnet'), + deployment: (import 'deployment.libsonnet'), + deploymentCondition: (import 'deploymentCondition.libsonnet'), + deploymentSpec: (import 'deploymentSpec.libsonnet'), + deploymentStatus: (import 'deploymentStatus.libsonnet'), + deploymentStrategy: (import 'deploymentStrategy.libsonnet'), + replicaSet: (import 'replicaSet.libsonnet'), + replicaSetCondition: (import 'replicaSetCondition.libsonnet'), + replicaSetSpec: (import 'replicaSetSpec.libsonnet'), + replicaSetStatus: (import 'replicaSetStatus.libsonnet'), + rollingUpdateDaemonSet: (import 'rollingUpdateDaemonSet.libsonnet'), + rollingUpdateDeployment: (import 'rollingUpdateDeployment.libsonnet'), + rollingUpdateStatefulSetStrategy: (import 'rollingUpdateStatefulSetStrategy.libsonnet'), + statefulSet: (import 'statefulSet.libsonnet'), + statefulSetCondition: (import 'statefulSetCondition.libsonnet'), + statefulSetSpec: (import 'statefulSetSpec.libsonnet'), + statefulSetStatus: (import 'statefulSetStatus.libsonnet'), + statefulSetUpdateStrategy: (import 'statefulSetUpdateStrategy.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSet.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSet.libsonnet new file mode 100644 index 00000000000..591aa455cce --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSet.libsonnet @@ -0,0 +1,313 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicaSet', url='', help='ReplicaSet ensures that a specified number of pod replicas are running at any given time.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Replicaset', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apps/v1', + kind: 'ReplicaSet', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='ReplicaSetSpec is the specification of a ReplicaSet.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withReplicas':: d.fn(help='Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetCondition.libsonnet new file mode 100644 index 00000000000..de9f448b195 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicaSetCondition', url='', help='ReplicaSetCondition describes the state of a replica set at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='A human readable message indicating details about the transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="The reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of replica set condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetSpec.libsonnet new file mode 100644 index 00000000000..db26a5bb17a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetSpec.libsonnet @@ -0,0 +1,260 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicaSetSpec', url='', help='ReplicaSetSpec is the specification of a ReplicaSet.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { minReadySeconds: minReadySeconds }, + '#withReplicas':: d.fn(help='Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetStatus.libsonnet new file mode 100644 index 00000000000..d4249a73c49 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/replicaSetStatus.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicaSetStatus', url='', help='ReplicaSetStatus represents the current status of a ReplicaSet.'), + '#withAvailableReplicas':: d.fn(help='The number of available replicas (ready for at least minReadySeconds) for this replica set.', args=[d.arg(name='availableReplicas', type=d.T.integer)]), + withAvailableReplicas(availableReplicas): { availableReplicas: availableReplicas }, + '#withConditions':: d.fn(help="Represents the latest available observations of a replica set's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a replica set's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withFullyLabeledReplicas':: d.fn(help='The number of pods that have labels matching the labels of the pod template of the replicaset.', args=[d.arg(name='fullyLabeledReplicas', type=d.T.integer)]), + withFullyLabeledReplicas(fullyLabeledReplicas): { fullyLabeledReplicas: fullyLabeledReplicas }, + '#withObservedGeneration':: d.fn(help='ObservedGeneration reflects the generation of the most recently observed ReplicaSet.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#withReadyReplicas':: d.fn(help='The number of ready replicas for this replica set.', args=[d.arg(name='readyReplicas', type=d.T.integer)]), + withReadyReplicas(readyReplicas): { readyReplicas: readyReplicas }, + '#withReplicas':: d.fn(help='Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDaemonSet.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDaemonSet.libsonnet new file mode 100644 index 00000000000..766aafdf0c4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDaemonSet.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rollingUpdateDaemonSet', url='', help='Spec to control the desired behavior of daemon set rolling update.'), + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { maxUnavailable: maxUnavailable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDeployment.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDeployment.libsonnet new file mode 100644 index 00000000000..113e3e3befa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateDeployment.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rollingUpdateDeployment', url='', help='Spec to control the desired behavior of rolling update.'), + '#withMaxSurge':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxSurge', type=d.T.string)]), + withMaxSurge(maxSurge): { maxSurge: maxSurge }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { maxUnavailable: maxUnavailable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateStatefulSetStrategy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateStatefulSetStrategy.libsonnet new file mode 100644 index 00000000000..074c1fe8a50 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/rollingUpdateStatefulSetStrategy.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rollingUpdateStatefulSetStrategy', url='', help='RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType.'), + '#withPartition':: d.fn(help='Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0.', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { partition: partition }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSet.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSet.libsonnet new file mode 100644 index 00000000000..90a7ba60196 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSet.libsonnet @@ -0,0 +1,331 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='statefulSet', url='', help='StatefulSet represents a set of pods with consistent identities. Identities are defined as:\n - Network: A single stable DNS and hostname.\n - Storage: As many VolumeClaims as requested.\nThe StatefulSet guarantees that a given network identity will always map to the same storage identity.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Statefulset', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'apps/v1', + kind: 'StatefulSet', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='A StatefulSetSpec is the specification of a StatefulSet.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#updateStrategy':: d.obj(help='StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy.'), + updateStrategy: { + '#rollingUpdate':: d.obj(help='RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType.'), + rollingUpdate: { + '#withPartition':: d.fn(help='Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0.', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { updateStrategy+: { rollingUpdate+: { partition: partition } } } }, + }, + '#withType':: d.fn(help='Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { updateStrategy+: { type: type } } }, + }, + '#withPodManagementPolicy':: d.fn(help='podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once.', args=[d.arg(name='podManagementPolicy', type=d.T.string)]), + withPodManagementPolicy(podManagementPolicy): { spec+: { podManagementPolicy: podManagementPolicy } }, + '#withReplicas':: d.fn(help='replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + '#withRevisionHistoryLimit':: d.fn(help="revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10.", args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { spec+: { revisionHistoryLimit: revisionHistoryLimit } }, + '#withServiceName':: d.fn(help='serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { spec+: { serviceName: serviceName } }, + '#withVolumeClaimTemplates':: d.fn(help='volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name.', args=[d.arg(name='volumeClaimTemplates', type=d.T.array)]), + withVolumeClaimTemplates(volumeClaimTemplates): { spec+: { volumeClaimTemplates: if std.isArray(v=volumeClaimTemplates) then volumeClaimTemplates else [volumeClaimTemplates] } }, + '#withVolumeClaimTemplatesMixin':: d.fn(help='volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeClaimTemplates', type=d.T.array)]), + withVolumeClaimTemplatesMixin(volumeClaimTemplates): { spec+: { volumeClaimTemplates+: if std.isArray(v=volumeClaimTemplates) then volumeClaimTemplates else [volumeClaimTemplates] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetCondition.libsonnet new file mode 100644 index 00000000000..f2bce71b224 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='statefulSetCondition', url='', help='StatefulSetCondition describes the state of a statefulset at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='A human readable message indicating details about the transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="The reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of statefulset condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetSpec.libsonnet new file mode 100644 index 00000000000..ad504930e03 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetSpec.libsonnet @@ -0,0 +1,278 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='statefulSetSpec', url='', help='A StatefulSetSpec is the specification of a StatefulSet.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#updateStrategy':: d.obj(help='StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy.'), + updateStrategy: { + '#rollingUpdate':: d.obj(help='RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType.'), + rollingUpdate: { + '#withPartition':: d.fn(help='Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0.', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { updateStrategy+: { rollingUpdate+: { partition: partition } } }, + }, + '#withType':: d.fn(help='Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { updateStrategy+: { type: type } }, + }, + '#withPodManagementPolicy':: d.fn(help='podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once.', args=[d.arg(name='podManagementPolicy', type=d.T.string)]), + withPodManagementPolicy(podManagementPolicy): { podManagementPolicy: podManagementPolicy }, + '#withReplicas':: d.fn(help='replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withRevisionHistoryLimit':: d.fn(help="revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10.", args=[d.arg(name='revisionHistoryLimit', type=d.T.integer)]), + withRevisionHistoryLimit(revisionHistoryLimit): { revisionHistoryLimit: revisionHistoryLimit }, + '#withServiceName':: d.fn(help='serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { serviceName: serviceName }, + '#withVolumeClaimTemplates':: d.fn(help='volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name.', args=[d.arg(name='volumeClaimTemplates', type=d.T.array)]), + withVolumeClaimTemplates(volumeClaimTemplates): { volumeClaimTemplates: if std.isArray(v=volumeClaimTemplates) then volumeClaimTemplates else [volumeClaimTemplates] }, + '#withVolumeClaimTemplatesMixin':: d.fn(help='volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeClaimTemplates', type=d.T.array)]), + withVolumeClaimTemplatesMixin(volumeClaimTemplates): { volumeClaimTemplates+: if std.isArray(v=volumeClaimTemplates) then volumeClaimTemplates else [volumeClaimTemplates] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetStatus.libsonnet new file mode 100644 index 00000000000..8f5c4b7e407 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetStatus.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='statefulSetStatus', url='', help='StatefulSetStatus represents the current state of a StatefulSet.'), + '#withCollisionCount':: d.fn(help='collisionCount is the count of hash collisions for the StatefulSet. The StatefulSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision.', args=[d.arg(name='collisionCount', type=d.T.integer)]), + withCollisionCount(collisionCount): { collisionCount: collisionCount }, + '#withConditions':: d.fn(help="Represents the latest available observations of a statefulset's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a statefulset's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withCurrentReplicas':: d.fn(help='currentReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by currentRevision.', args=[d.arg(name='currentReplicas', type=d.T.integer)]), + withCurrentReplicas(currentReplicas): { currentReplicas: currentReplicas }, + '#withCurrentRevision':: d.fn(help='currentRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [0,currentReplicas).', args=[d.arg(name='currentRevision', type=d.T.string)]), + withCurrentRevision(currentRevision): { currentRevision: currentRevision }, + '#withObservedGeneration':: d.fn(help="observedGeneration is the most recent generation observed for this StatefulSet. It corresponds to the StatefulSet's generation, which is updated on mutation by the API Server.", args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#withReadyReplicas':: d.fn(help='readyReplicas is the number of Pods created by the StatefulSet controller that have a Ready Condition.', args=[d.arg(name='readyReplicas', type=d.T.integer)]), + withReadyReplicas(readyReplicas): { readyReplicas: readyReplicas }, + '#withReplicas':: d.fn(help='replicas is the number of Pods created by the StatefulSet controller.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withUpdateRevision':: d.fn(help='updateRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [replicas-updatedReplicas,replicas)', args=[d.arg(name='updateRevision', type=d.T.string)]), + withUpdateRevision(updateRevision): { updateRevision: updateRevision }, + '#withUpdatedReplicas':: d.fn(help='updatedReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by updateRevision.', args=[d.arg(name='updatedReplicas', type=d.T.integer)]), + withUpdatedReplicas(updatedReplicas): { updatedReplicas: updatedReplicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetUpdateStrategy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetUpdateStrategy.libsonnet new file mode 100644 index 00000000000..0b0cbfe2ac1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/apps/v1/statefulSetUpdateStrategy.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='statefulSetUpdateStrategy', url='', help='StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy.'), + '#rollingUpdate':: d.obj(help='RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType.'), + rollingUpdate: { + '#withPartition':: d.fn(help='Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0.', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { rollingUpdate+: { partition: partition } }, + }, + '#withType':: d.fn(help='Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/main.libsonnet new file mode 100644 index 00000000000..79f614c9bef --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='authentication', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/boundObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/boundObjectReference.libsonnet new file mode 100644 index 00000000000..574787c0447 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/boundObjectReference.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='boundObjectReference', url='', help='BoundObjectReference is a reference to an object that a token is bound to.'), + '#withKind':: d.fn(help="Kind of the referent. Valid kinds are 'Pod' and 'Secret'.", args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the referent.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withUid':: d.fn(help='UID of the referent.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/main.libsonnet new file mode 100644 index 00000000000..05cedc80c76 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/main.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + boundObjectReference: (import 'boundObjectReference.libsonnet'), + tokenRequest: (import 'tokenRequest.libsonnet'), + tokenRequestSpec: (import 'tokenRequestSpec.libsonnet'), + tokenRequestStatus: (import 'tokenRequestStatus.libsonnet'), + tokenReview: (import 'tokenReview.libsonnet'), + tokenReviewSpec: (import 'tokenReviewSpec.libsonnet'), + tokenReviewStatus: (import 'tokenReviewStatus.libsonnet'), + userInfo: (import 'userInfo.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequest.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequest.libsonnet new file mode 100644 index 00000000000..f3be4ae3959 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequest.libsonnet @@ -0,0 +1,76 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenRequest', url='', help='TokenRequest requests a token for a given service account.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Tokenrequest', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authentication.k8s.io/v1', + kind: 'TokenRequest', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='TokenRequestSpec contains client provided parameters of a token request.'), + spec: { + '#boundObjectRef':: d.obj(help='BoundObjectReference is a reference to an object that a token is bound to.'), + boundObjectRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { boundObjectRef+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help="Kind of the referent. Valid kinds are 'Pod' and 'Secret'.", args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { boundObjectRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { boundObjectRef+: { name: name } } }, + '#withUid':: d.fn(help='UID of the referent.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { boundObjectRef+: { uid: uid } } }, + }, + '#withAudiences':: d.fn(help='Audiences are the intendend audiences of the token. A recipient of a token must identitfy themself with an identifier in the list of audiences of the token, and otherwise should reject the token. A token issued for multiple audiences may be used to authenticate against any of the audiences listed but implies a high degree of trust between the target audiences.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { spec+: { audiences: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withAudiencesMixin':: d.fn(help='Audiences are the intendend audiences of the token. A recipient of a token must identitfy themself with an identifier in the list of audiences of the token, and otherwise should reject the token. A token issued for multiple audiences may be used to authenticate against any of the audiences listed but implies a high degree of trust between the target audiences.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { spec+: { audiences+: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withExpirationSeconds':: d.fn(help="ExpirationSeconds is the requested duration of validity of the request. The token issuer may return a token with a different validity duration so a client needs to check the 'expiration' field in a response.", args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { spec+: { expirationSeconds: expirationSeconds } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestSpec.libsonnet new file mode 100644 index 00000000000..5fa0c6b77cb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestSpec.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenRequestSpec', url='', help='TokenRequestSpec contains client provided parameters of a token request.'), + '#boundObjectRef':: d.obj(help='BoundObjectReference is a reference to an object that a token is bound to.'), + boundObjectRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { boundObjectRef+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help="Kind of the referent. Valid kinds are 'Pod' and 'Secret'.", args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { boundObjectRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { boundObjectRef+: { name: name } }, + '#withUid':: d.fn(help='UID of the referent.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { boundObjectRef+: { uid: uid } }, + }, + '#withAudiences':: d.fn(help='Audiences are the intendend audiences of the token. A recipient of a token must identitfy themself with an identifier in the list of audiences of the token, and otherwise should reject the token. A token issued for multiple audiences may be used to authenticate against any of the audiences listed but implies a high degree of trust between the target audiences.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { audiences: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAudiencesMixin':: d.fn(help='Audiences are the intendend audiences of the token. A recipient of a token must identitfy themself with an identifier in the list of audiences of the token, and otherwise should reject the token. A token issued for multiple audiences may be used to authenticate against any of the audiences listed but implies a high degree of trust between the target audiences.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { audiences+: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withExpirationSeconds':: d.fn(help="ExpirationSeconds is the requested duration of validity of the request. The token issuer may return a token with a different validity duration so a client needs to check the 'expiration' field in a response.", args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { expirationSeconds: expirationSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestStatus.libsonnet new file mode 100644 index 00000000000..d26b07133c4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenRequestStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenRequestStatus', url='', help='TokenRequestStatus is the result of a token request.'), + '#withExpirationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='expirationTimestamp', type=d.T.string)]), + withExpirationTimestamp(expirationTimestamp): { expirationTimestamp: expirationTimestamp }, + '#withToken':: d.fn(help='Token is the opaque bearer token.', args=[d.arg(name='token', type=d.T.string)]), + withToken(token): { token: token }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReview.libsonnet new file mode 100644 index 00000000000..2d285978743 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReview.libsonnet @@ -0,0 +1,65 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReview', url='', help='TokenReview attempts to authenticate a token to a known user. Note: TokenReview requests may be cached by the webhook token authenticator plugin in the kube-apiserver.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Tokenreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authentication.k8s.io/v1', + kind: 'TokenReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='TokenReviewSpec is a description of the token authentication request.'), + spec: { + '#withAudiences':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { spec+: { audiences: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withAudiencesMixin':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { spec+: { audiences+: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withToken':: d.fn(help='Token is the opaque bearer token.', args=[d.arg(name='token', type=d.T.string)]), + withToken(token): { spec+: { token: token } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewSpec.libsonnet new file mode 100644 index 00000000000..0e0e0b8e3be --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewSpec.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReviewSpec', url='', help='TokenReviewSpec is a description of the token authentication request.'), + '#withAudiences':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { audiences: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAudiencesMixin':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { audiences+: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withToken':: d.fn(help='Token is the opaque bearer token.', args=[d.arg(name='token', type=d.T.string)]), + withToken(token): { token: token }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewStatus.libsonnet new file mode 100644 index 00000000000..3a816c73958 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/tokenReviewStatus.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReviewStatus', url='', help='TokenReviewStatus is the result of the token authentication request.'), + '#user':: d.obj(help='UserInfo holds the information about the user needed to implement the user.Info interface.'), + user: { + '#withExtra':: d.fn(help='Any additional information provided by the authenticator.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { user+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Any additional information provided by the authenticator.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { user+: { extra+: extra } }, + '#withGroups':: d.fn(help='The names of groups this user is a part of.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { user+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help='The names of groups this user is a part of.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { user+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withUid':: d.fn(help='A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { user+: { uid: uid } }, + '#withUsername':: d.fn(help='The name that uniquely identifies this user among all active users.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { user+: { username: username } }, + }, + '#withAudiences':: d.fn(help="Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is 'true', the token is valid against the audience of the Kubernetes API server.", args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { audiences: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAudiencesMixin':: d.fn(help="Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is 'true', the token is valid against the audience of the Kubernetes API server.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { audiences+: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAuthenticated':: d.fn(help='Authenticated indicates that the token was associated with a known user.', args=[d.arg(name='authenticated', type=d.T.boolean)]), + withAuthenticated(authenticated): { authenticated: authenticated }, + '#withError':: d.fn(help="Error indicates that the token couldn't be checked", args=[d.arg(name='err', type=d.T.string)]), + withError(err): { 'error': err }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/userInfo.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/userInfo.libsonnet new file mode 100644 index 00000000000..93ed829427f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1/userInfo.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='userInfo', url='', help='UserInfo holds the information about the user needed to implement the user.Info interface.'), + '#withExtra':: d.fn(help='Any additional information provided by the authenticator.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='Any additional information provided by the authenticator.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroups':: d.fn(help='The names of groups this user is a part of.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { groups: if std.isArray(v=groups) then groups else [groups] }, + '#withGroupsMixin':: d.fn(help='The names of groups this user is a part of.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { groups+: if std.isArray(v=groups) then groups else [groups] }, + '#withUid':: d.fn(help='A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUsername':: d.fn(help='The name that uniquely identifies this user among all active users.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { username: username }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/main.libsonnet new file mode 100644 index 00000000000..5c9a07f9a3c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + tokenReview: (import 'tokenReview.libsonnet'), + tokenReviewSpec: (import 'tokenReviewSpec.libsonnet'), + tokenReviewStatus: (import 'tokenReviewStatus.libsonnet'), + userInfo: (import 'userInfo.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReview.libsonnet new file mode 100644 index 00000000000..18cb1ee0f2d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReview.libsonnet @@ -0,0 +1,65 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReview', url='', help='TokenReview attempts to authenticate a token to a known user. Note: TokenReview requests may be cached by the webhook token authenticator plugin in the kube-apiserver.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Tokenreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authentication.k8s.io/v1beta1', + kind: 'TokenReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='TokenReviewSpec is a description of the token authentication request.'), + spec: { + '#withAudiences':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { spec+: { audiences: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withAudiencesMixin':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { spec+: { audiences+: if std.isArray(v=audiences) then audiences else [audiences] } }, + '#withToken':: d.fn(help='Token is the opaque bearer token.', args=[d.arg(name='token', type=d.T.string)]), + withToken(token): { spec+: { token: token } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewSpec.libsonnet new file mode 100644 index 00000000000..0e0e0b8e3be --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewSpec.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReviewSpec', url='', help='TokenReviewSpec is a description of the token authentication request.'), + '#withAudiences':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { audiences: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAudiencesMixin':: d.fn(help='Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { audiences+: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withToken':: d.fn(help='Token is the opaque bearer token.', args=[d.arg(name='token', type=d.T.string)]), + withToken(token): { token: token }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewStatus.libsonnet new file mode 100644 index 00000000000..3a816c73958 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/tokenReviewStatus.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tokenReviewStatus', url='', help='TokenReviewStatus is the result of the token authentication request.'), + '#user':: d.obj(help='UserInfo holds the information about the user needed to implement the user.Info interface.'), + user: { + '#withExtra':: d.fn(help='Any additional information provided by the authenticator.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { user+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Any additional information provided by the authenticator.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { user+: { extra+: extra } }, + '#withGroups':: d.fn(help='The names of groups this user is a part of.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { user+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help='The names of groups this user is a part of.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { user+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withUid':: d.fn(help='A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { user+: { uid: uid } }, + '#withUsername':: d.fn(help='The name that uniquely identifies this user among all active users.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { user+: { username: username } }, + }, + '#withAudiences':: d.fn(help="Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is 'true', the token is valid against the audience of the Kubernetes API server.", args=[d.arg(name='audiences', type=d.T.array)]), + withAudiences(audiences): { audiences: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAudiencesMixin':: d.fn(help="Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is 'true', the token is valid against the audience of the Kubernetes API server.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='audiences', type=d.T.array)]), + withAudiencesMixin(audiences): { audiences+: if std.isArray(v=audiences) then audiences else [audiences] }, + '#withAuthenticated':: d.fn(help='Authenticated indicates that the token was associated with a known user.', args=[d.arg(name='authenticated', type=d.T.boolean)]), + withAuthenticated(authenticated): { authenticated: authenticated }, + '#withError':: d.fn(help="Error indicates that the token couldn't be checked", args=[d.arg(name='err', type=d.T.string)]), + withError(err): { 'error': err }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/userInfo.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/userInfo.libsonnet new file mode 100644 index 00000000000..93ed829427f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authentication/v1beta1/userInfo.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='userInfo', url='', help='UserInfo holds the information about the user needed to implement the user.Info interface.'), + '#withExtra':: d.fn(help='Any additional information provided by the authenticator.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='Any additional information provided by the authenticator.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroups':: d.fn(help='The names of groups this user is a part of.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { groups: if std.isArray(v=groups) then groups else [groups] }, + '#withGroupsMixin':: d.fn(help='The names of groups this user is a part of.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { groups+: if std.isArray(v=groups) then groups else [groups] }, + '#withUid':: d.fn(help='A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUsername':: d.fn(help='The name that uniquely identifies this user among all active users.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { username: username }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/main.libsonnet new file mode 100644 index 00000000000..c56c20bb553 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='authorization', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/localSubjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/localSubjectAccessReview.libsonnet new file mode 100644 index 00000000000..53699d1dc21 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/localSubjectAccessReview.libsonnet @@ -0,0 +1,95 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='localSubjectAccessReview', url='', help='LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. Having a namespace scoped resource makes it much easier to grant namespace scoped policy that includes permissions checking.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Localsubjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1', + kind: 'LocalSubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroups':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { spec+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { spec+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Groups', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { user: user } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/main.libsonnet new file mode 100644 index 00000000000..61af66b02b4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/main.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + localSubjectAccessReview: (import 'localSubjectAccessReview.libsonnet'), + nonResourceAttributes: (import 'nonResourceAttributes.libsonnet'), + nonResourceRule: (import 'nonResourceRule.libsonnet'), + resourceAttributes: (import 'resourceAttributes.libsonnet'), + resourceRule: (import 'resourceRule.libsonnet'), + selfSubjectAccessReview: (import 'selfSubjectAccessReview.libsonnet'), + selfSubjectAccessReviewSpec: (import 'selfSubjectAccessReviewSpec.libsonnet'), + selfSubjectRulesReview: (import 'selfSubjectRulesReview.libsonnet'), + selfSubjectRulesReviewSpec: (import 'selfSubjectRulesReviewSpec.libsonnet'), + subjectAccessReview: (import 'subjectAccessReview.libsonnet'), + subjectAccessReviewSpec: (import 'subjectAccessReviewSpec.libsonnet'), + subjectAccessReviewStatus: (import 'subjectAccessReviewStatus.libsonnet'), + subjectRulesReviewStatus: (import 'subjectRulesReviewStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceAttributes.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceAttributes.libsonnet new file mode 100644 index 00000000000..1d6844be6b3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceAttributes.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nonResourceAttributes', url='', help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { verb: verb }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceRule.libsonnet new file mode 100644 index 00000000000..b138fef63dd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/nonResourceRule.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nonResourceRule', url='', help='NonResourceRule holds information that describes a rule for the non-resource'), + '#withNonResourceURLs':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withVerbs':: d.fn(help='Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceAttributes.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceAttributes.libsonnet new file mode 100644 index 00000000000..26234c41d70 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceAttributes.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceAttributes', url='', help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resource: resource }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { subresource: subresource }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { verb: verb }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { version: version }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceRule.libsonnet new file mode 100644 index 00000000000..70cdabcd5aa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/resourceRule.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceRule', url='', help="ResourceRule is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete."), + '#withApiGroups':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withResourceNames':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all.', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNames(resourceNames): { resourceNames: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResourceNamesMixin':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNamesMixin(resourceNames): { resourceNames+: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResources':: d.fn(help="Resources is a list of resources this rule applies to. '*' means all in the specified apiGroups.\n '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.", args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help="Resources is a list of resources this rule applies to. '*' means all in the specified apiGroups.\n '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReview.libsonnet new file mode 100644 index 00000000000..769b09c9e68 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReview.libsonnet @@ -0,0 +1,83 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectAccessReview', url='', help='SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a spec.namespace means "in all namespaces". Self is a special case, because users should always be able to check whether they can perform an action'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Selfsubjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1', + kind: 'SelfSubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReviewSpec.libsonnet new file mode 100644 index 00000000000..6462f5c5c18 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectAccessReviewSpec.libsonnet @@ -0,0 +1,30 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectAccessReviewSpec', url='', help='SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nonResourceAttributes+: { path: path } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { nonResourceAttributes+: { verb: verb } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { resourceAttributes+: { group: group } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resourceAttributes+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { resourceAttributes+: { namespace: namespace } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceAttributes+: { resource: resource } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { resourceAttributes+: { subresource: subresource } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { resourceAttributes+: { verb: verb } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { resourceAttributes+: { version: version } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReview.libsonnet new file mode 100644 index 00000000000..a11f3d1009d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReview.libsonnet @@ -0,0 +1,61 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectRulesReview', url='', help="SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. The returned list of actions may be incomplete depending on the server's authorization mode, and any errors experienced during the evaluation. SelfSubjectRulesReview should be used by UIs to show/hide actions, or to quickly let an end user reason about their permissions. It should NOT Be used by external systems to drive authorization decisions as this raises confused deputy, cache lifetime/revocation, and correctness concerns. SubjectAccessReview, and LocalAccessReview are the correct way to defer authorization decisions to the API server."), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Selfsubjectrulesreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1', + kind: 'SelfSubjectRulesReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help=''), + spec: { + '#withNamespace':: d.fn(help='Namespace to evaluate rules for. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { namespace: namespace } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReviewSpec.libsonnet new file mode 100644 index 00000000000..e71b052da1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/selfSubjectRulesReviewSpec.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectRulesReviewSpec', url='', help=''), + '#withNamespace':: d.fn(help='Namespace to evaluate rules for. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReview.libsonnet new file mode 100644 index 00000000000..6b5950055f1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReview.libsonnet @@ -0,0 +1,95 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReview', url='', help='SubjectAccessReview checks whether or not a user or group can perform an action.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Subjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1', + kind: 'SubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroups':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { spec+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { spec+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Groups', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { user: user } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewSpec.libsonnet new file mode 100644 index 00000000000..7f57a47fb2f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewSpec.libsonnet @@ -0,0 +1,42 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReviewSpec', url='', help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nonResourceAttributes+: { path: path } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { nonResourceAttributes+: { verb: verb } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { resourceAttributes+: { group: group } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resourceAttributes+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { resourceAttributes+: { namespace: namespace } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceAttributes+: { resource: resource } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { resourceAttributes+: { subresource: subresource } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { resourceAttributes+: { verb: verb } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { resourceAttributes+: { version: version } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroups':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { groups: if std.isArray(v=groups) then groups else [groups] }, + '#withGroupsMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { groups+: if std.isArray(v=groups) then groups else [groups] }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Groups', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewStatus.libsonnet new file mode 100644 index 00000000000..e8b77e1031c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectAccessReviewStatus.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReviewStatus', url='', help='SubjectAccessReviewStatus'), + '#withAllowed':: d.fn(help='Allowed is required. True if the action would be allowed, false otherwise.', args=[d.arg(name='allowed', type=d.T.boolean)]), + withAllowed(allowed): { allowed: allowed }, + '#withDenied':: d.fn(help='Denied is optional. True if the action would be denied, otherwise false. If both allowed is false and denied is false, then the authorizer has no opinion on whether to authorize the action. Denied may not be true if Allowed is true.', args=[d.arg(name='denied', type=d.T.boolean)]), + withDenied(denied): { denied: denied }, + '#withEvaluationError':: d.fn(help='EvaluationError is an indication that some error occurred during the authorization check. It is entirely possible to get an error and be able to continue determine authorization status in spite of it. For instance, RBAC can be missing a role, but enough roles are still present and bound to reason about the request.', args=[d.arg(name='evaluationError', type=d.T.string)]), + withEvaluationError(evaluationError): { evaluationError: evaluationError }, + '#withReason':: d.fn(help='Reason is optional. It indicates why a request was allowed or denied.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectRulesReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectRulesReviewStatus.libsonnet new file mode 100644 index 00000000000..9dd07031d01 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1/subjectRulesReviewStatus.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectRulesReviewStatus', url='', help="SubjectRulesReviewStatus contains the result of a rules check. This check can be incomplete depending on the set of authorizers the server is configured with and any errors experienced during evaluation. Because authorization rules are additive, if a rule appears in a list it's safe to assume the subject has that permission, even if that list is incomplete."), + '#withEvaluationError':: d.fn(help="EvaluationError can appear in combination with Rules. It indicates an error occurred during rule evaluation, such as an authorizer that doesn't support rule evaluation, and that ResourceRules and/or NonResourceRules may be incomplete.", args=[d.arg(name='evaluationError', type=d.T.string)]), + withEvaluationError(evaluationError): { evaluationError: evaluationError }, + '#withIncomplete':: d.fn(help="Incomplete is true when the rules returned by this call are incomplete. This is most commonly encountered when an authorizer, such as an external authorizer, doesn't support rules evaluation.", args=[d.arg(name='incomplete', type=d.T.boolean)]), + withIncomplete(incomplete): { incomplete: incomplete }, + '#withNonResourceRules':: d.fn(help="NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.", args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRules(nonResourceRules): { nonResourceRules: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withNonResourceRulesMixin':: d.fn(help="NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRulesMixin(nonResourceRules): { nonResourceRules+: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withResourceRules':: d.fn(help="ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.", args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRules(resourceRules): { resourceRules: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#withResourceRulesMixin':: d.fn(help="ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRulesMixin(resourceRules): { resourceRules+: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/localSubjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/localSubjectAccessReview.libsonnet new file mode 100644 index 00000000000..ce47e09014f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/localSubjectAccessReview.libsonnet @@ -0,0 +1,95 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='localSubjectAccessReview', url='', help='LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. Having a namespace scoped resource makes it much easier to grant namespace scoped policy that includes permissions checking.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Localsubjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1beta1', + kind: 'LocalSubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroup':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='group', type=d.T.array)]), + withGroup(group): { spec+: { group: if std.isArray(v=group) then group else [group] } }, + '#withGroupMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='group', type=d.T.array)]), + withGroupMixin(group): { spec+: { group+: if std.isArray(v=group) then group else [group] } }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Group', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { user: user } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/main.libsonnet new file mode 100644 index 00000000000..58919359023 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/main.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + localSubjectAccessReview: (import 'localSubjectAccessReview.libsonnet'), + nonResourceAttributes: (import 'nonResourceAttributes.libsonnet'), + nonResourceRule: (import 'nonResourceRule.libsonnet'), + resourceAttributes: (import 'resourceAttributes.libsonnet'), + resourceRule: (import 'resourceRule.libsonnet'), + selfSubjectAccessReview: (import 'selfSubjectAccessReview.libsonnet'), + selfSubjectAccessReviewSpec: (import 'selfSubjectAccessReviewSpec.libsonnet'), + selfSubjectRulesReview: (import 'selfSubjectRulesReview.libsonnet'), + selfSubjectRulesReviewSpec: (import 'selfSubjectRulesReviewSpec.libsonnet'), + subjectAccessReview: (import 'subjectAccessReview.libsonnet'), + subjectAccessReviewSpec: (import 'subjectAccessReviewSpec.libsonnet'), + subjectAccessReviewStatus: (import 'subjectAccessReviewStatus.libsonnet'), + subjectRulesReviewStatus: (import 'subjectRulesReviewStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceAttributes.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceAttributes.libsonnet new file mode 100644 index 00000000000..1d6844be6b3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceAttributes.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nonResourceAttributes', url='', help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { verb: verb }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceRule.libsonnet new file mode 100644 index 00000000000..b138fef63dd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/nonResourceRule.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nonResourceRule', url='', help='NonResourceRule holds information that describes a rule for the non-resource'), + '#withNonResourceURLs':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withVerbs':: d.fn(help='Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceAttributes.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceAttributes.libsonnet new file mode 100644 index 00000000000..26234c41d70 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceAttributes.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceAttributes', url='', help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resource: resource }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { subresource: subresource }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { verb: verb }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { version: version }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceRule.libsonnet new file mode 100644 index 00000000000..70cdabcd5aa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/resourceRule.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceRule', url='', help="ResourceRule is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete."), + '#withApiGroups':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withResourceNames':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all.', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNames(resourceNames): { resourceNames: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResourceNamesMixin':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNamesMixin(resourceNames): { resourceNames+: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResources':: d.fn(help="Resources is a list of resources this rule applies to. '*' means all in the specified apiGroups.\n '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.", args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help="Resources is a list of resources this rule applies to. '*' means all in the specified apiGroups.\n '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReview.libsonnet new file mode 100644 index 00000000000..3b7aabae75d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReview.libsonnet @@ -0,0 +1,83 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectAccessReview', url='', help='SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a spec.namespace means "in all namespaces". Self is a special case, because users should always be able to check whether they can perform an action'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Selfsubjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1beta1', + kind: 'SelfSubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReviewSpec.libsonnet new file mode 100644 index 00000000000..6462f5c5c18 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectAccessReviewSpec.libsonnet @@ -0,0 +1,30 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectAccessReviewSpec', url='', help='SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nonResourceAttributes+: { path: path } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { nonResourceAttributes+: { verb: verb } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { resourceAttributes+: { group: group } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resourceAttributes+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { resourceAttributes+: { namespace: namespace } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceAttributes+: { resource: resource } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { resourceAttributes+: { subresource: subresource } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { resourceAttributes+: { verb: verb } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { resourceAttributes+: { version: version } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReview.libsonnet new file mode 100644 index 00000000000..195f074152f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReview.libsonnet @@ -0,0 +1,61 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectRulesReview', url='', help="SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. The returned list of actions may be incomplete depending on the server's authorization mode, and any errors experienced during the evaluation. SelfSubjectRulesReview should be used by UIs to show/hide actions, or to quickly let an end user reason about their permissions. It should NOT Be used by external systems to drive authorization decisions as this raises confused deputy, cache lifetime/revocation, and correctness concerns. SubjectAccessReview, and LocalAccessReview are the correct way to defer authorization decisions to the API server."), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Selfsubjectrulesreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1beta1', + kind: 'SelfSubjectRulesReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help=''), + spec: { + '#withNamespace':: d.fn(help='Namespace to evaluate rules for. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { namespace: namespace } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReviewSpec.libsonnet new file mode 100644 index 00000000000..e71b052da1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/selfSubjectRulesReviewSpec.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='selfSubjectRulesReviewSpec', url='', help=''), + '#withNamespace':: d.fn(help='Namespace to evaluate rules for. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReview.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReview.libsonnet new file mode 100644 index 00000000000..eed66cb82aa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReview.libsonnet @@ -0,0 +1,95 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReview', url='', help='SubjectAccessReview checks whether or not a user or group can perform an action.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Subjectaccessreview', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'authorization.k8s.io/v1beta1', + kind: 'SubjectAccessReview', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + spec: { + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nonResourceAttributes+: { path: path } } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { nonResourceAttributes+: { verb: verb } } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { resourceAttributes+: { group: group } } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { resourceAttributes+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { resourceAttributes+: { namespace: namespace } } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { spec+: { resourceAttributes+: { resource: resource } } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { spec+: { resourceAttributes+: { subresource: subresource } } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { spec+: { resourceAttributes+: { verb: verb } } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { resourceAttributes+: { version: version } } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroup':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='group', type=d.T.array)]), + withGroup(group): { spec+: { group: if std.isArray(v=group) then group else [group] } }, + '#withGroupMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='group', type=d.T.array)]), + withGroupMixin(group): { spec+: { group+: if std.isArray(v=group) then group else [group] } }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Group', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { user: user } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewSpec.libsonnet new file mode 100644 index 00000000000..24f729e0a6e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewSpec.libsonnet @@ -0,0 +1,42 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReviewSpec', url='', help='SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set'), + '#nonResourceAttributes':: d.obj(help='NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface'), + nonResourceAttributes: { + '#withPath':: d.fn(help='Path is the URL path of the request', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nonResourceAttributes+: { path: path } }, + '#withVerb':: d.fn(help='Verb is the standard HTTP verb', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { nonResourceAttributes+: { verb: verb } }, + }, + '#resourceAttributes':: d.obj(help='ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface'), + resourceAttributes: { + '#withGroup':: d.fn(help='Group is the API Group of the Resource. "*" means all.', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { resourceAttributes+: { group: group } }, + '#withName':: d.fn(help='Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resourceAttributes+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { resourceAttributes+: { namespace: namespace } }, + '#withResource':: d.fn(help='Resource is one of the existing resource types. "*" means all.', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceAttributes+: { resource: resource } }, + '#withSubresource':: d.fn(help='Subresource is one of the existing resource types. "" means none.', args=[d.arg(name='subresource', type=d.T.string)]), + withSubresource(subresource): { resourceAttributes+: { subresource: subresource } }, + '#withVerb':: d.fn(help='Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all.', args=[d.arg(name='verb', type=d.T.string)]), + withVerb(verb): { resourceAttributes+: { verb: verb } }, + '#withVersion':: d.fn(help='Version is the API Version of the Resource. "*" means all.', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { resourceAttributes+: { version: version } }, + }, + '#withExtra':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroup':: d.fn(help="Groups is the groups you're testing for.", args=[d.arg(name='group', type=d.T.array)]), + withGroup(group): { group: if std.isArray(v=group) then group else [group] }, + '#withGroupMixin':: d.fn(help="Groups is the groups you're testing for.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='group', type=d.T.array)]), + withGroupMixin(group): { group+: if std.isArray(v=group) then group else [group] }, + '#withUid':: d.fn(help='UID information about the requesting user.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUser':: d.fn(help="User is the user you're testing for. If you specify 'User' but not 'Group', then is it interpreted as 'What if User were not a member of any groups", args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewStatus.libsonnet new file mode 100644 index 00000000000..e8b77e1031c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectAccessReviewStatus.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectAccessReviewStatus', url='', help='SubjectAccessReviewStatus'), + '#withAllowed':: d.fn(help='Allowed is required. True if the action would be allowed, false otherwise.', args=[d.arg(name='allowed', type=d.T.boolean)]), + withAllowed(allowed): { allowed: allowed }, + '#withDenied':: d.fn(help='Denied is optional. True if the action would be denied, otherwise false. If both allowed is false and denied is false, then the authorizer has no opinion on whether to authorize the action. Denied may not be true if Allowed is true.', args=[d.arg(name='denied', type=d.T.boolean)]), + withDenied(denied): { denied: denied }, + '#withEvaluationError':: d.fn(help='EvaluationError is an indication that some error occurred during the authorization check. It is entirely possible to get an error and be able to continue determine authorization status in spite of it. For instance, RBAC can be missing a role, but enough roles are still present and bound to reason about the request.', args=[d.arg(name='evaluationError', type=d.T.string)]), + withEvaluationError(evaluationError): { evaluationError: evaluationError }, + '#withReason':: d.fn(help='Reason is optional. It indicates why a request was allowed or denied.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectRulesReviewStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectRulesReviewStatus.libsonnet new file mode 100644 index 00000000000..9dd07031d01 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/authorization/v1beta1/subjectRulesReviewStatus.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subjectRulesReviewStatus', url='', help="SubjectRulesReviewStatus contains the result of a rules check. This check can be incomplete depending on the set of authorizers the server is configured with and any errors experienced during evaluation. Because authorization rules are additive, if a rule appears in a list it's safe to assume the subject has that permission, even if that list is incomplete."), + '#withEvaluationError':: d.fn(help="EvaluationError can appear in combination with Rules. It indicates an error occurred during rule evaluation, such as an authorizer that doesn't support rule evaluation, and that ResourceRules and/or NonResourceRules may be incomplete.", args=[d.arg(name='evaluationError', type=d.T.string)]), + withEvaluationError(evaluationError): { evaluationError: evaluationError }, + '#withIncomplete':: d.fn(help="Incomplete is true when the rules returned by this call are incomplete. This is most commonly encountered when an authorizer, such as an external authorizer, doesn't support rules evaluation.", args=[d.arg(name='incomplete', type=d.T.boolean)]), + withIncomplete(incomplete): { incomplete: incomplete }, + '#withNonResourceRules':: d.fn(help="NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.", args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRules(nonResourceRules): { nonResourceRules: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withNonResourceRulesMixin':: d.fn(help="NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRulesMixin(nonResourceRules): { nonResourceRules+: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withResourceRules':: d.fn(help="ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.", args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRules(resourceRules): { resourceRules: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#withResourceRulesMixin':: d.fn(help="ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRulesMixin(resourceRules): { resourceRules+: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/main.libsonnet new file mode 100644 index 00000000000..69a8e08c746 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='autoscaling', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v2beta1: (import 'v2beta1/main.libsonnet'), + v2beta2: (import 'v2beta2/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/crossVersionObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/crossVersionObjectReference.libsonnet new file mode 100644 index 00000000000..b5eb31cc2e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/crossVersionObjectReference.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='crossVersionObjectReference', url='', help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscaler.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscaler.libsonnet new file mode 100644 index 00000000000..f5fe4e0088b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscaler.libsonnet @@ -0,0 +1,74 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscaler', url='', help='configuration of a horizontal pod autoscaler.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Horizontalpodautoscaler', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'autoscaling/v1', + kind: 'HorizontalPodAutoscaler', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='specification of a horizontal pod autoscaler.'), + spec: { + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { scaleTargetRef+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { scaleTargetRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { scaleTargetRef+: { name: name } } }, + }, + '#withMaxReplicas':: d.fn(help='upper limit for the number of pods that can be set by the autoscaler; cannot be smaller than MinReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { spec+: { maxReplicas: maxReplicas } }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { spec+: { minReplicas: minReplicas } }, + '#withTargetCPUUtilizationPercentage':: d.fn(help='target average CPU utilization (represented as a percentage of requested CPU) over all the pods; if not specified the default autoscaling policy will be used.', args=[d.arg(name='targetCPUUtilizationPercentage', type=d.T.integer)]), + withTargetCPUUtilizationPercentage(targetCPUUtilizationPercentage): { spec+: { targetCPUUtilizationPercentage: targetCPUUtilizationPercentage } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerSpec.libsonnet new file mode 100644 index 00000000000..59c33f5ba08 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerSpec.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerSpec', url='', help='specification of a horizontal pod autoscaler.'), + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { scaleTargetRef+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { scaleTargetRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleTargetRef+: { name: name } }, + }, + '#withMaxReplicas':: d.fn(help='upper limit for the number of pods that can be set by the autoscaler; cannot be smaller than MinReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { maxReplicas: maxReplicas }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { minReplicas: minReplicas }, + '#withTargetCPUUtilizationPercentage':: d.fn(help='target average CPU utilization (represented as a percentage of requested CPU) over all the pods; if not specified the default autoscaling policy will be used.', args=[d.arg(name='targetCPUUtilizationPercentage', type=d.T.integer)]), + withTargetCPUUtilizationPercentage(targetCPUUtilizationPercentage): { targetCPUUtilizationPercentage: targetCPUUtilizationPercentage }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerStatus.libsonnet new file mode 100644 index 00000000000..c6c10684817 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/horizontalPodAutoscalerStatus.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerStatus', url='', help='current status of a horizontal pod autoscaler'), + '#withCurrentCPUUtilizationPercentage':: d.fn(help='current average CPU utilization over all pods, represented as a percentage of requested CPU, e.g. 70 means that an average pod is using now 70% of its requested CPU.', args=[d.arg(name='currentCPUUtilizationPercentage', type=d.T.integer)]), + withCurrentCPUUtilizationPercentage(currentCPUUtilizationPercentage): { currentCPUUtilizationPercentage: currentCPUUtilizationPercentage }, + '#withCurrentReplicas':: d.fn(help='current number of replicas of pods managed by this autoscaler.', args=[d.arg(name='currentReplicas', type=d.T.integer)]), + withCurrentReplicas(currentReplicas): { currentReplicas: currentReplicas }, + '#withDesiredReplicas':: d.fn(help='desired number of replicas of pods managed by this autoscaler.', args=[d.arg(name='desiredReplicas', type=d.T.integer)]), + withDesiredReplicas(desiredReplicas): { desiredReplicas: desiredReplicas }, + '#withLastScaleTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastScaleTime', type=d.T.string)]), + withLastScaleTime(lastScaleTime): { lastScaleTime: lastScaleTime }, + '#withObservedGeneration':: d.fn(help='most recent generation observed by this autoscaler.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/main.libsonnet new file mode 100644 index 00000000000..2e4434d770b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/main.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + crossVersionObjectReference: (import 'crossVersionObjectReference.libsonnet'), + horizontalPodAutoscaler: (import 'horizontalPodAutoscaler.libsonnet'), + horizontalPodAutoscalerSpec: (import 'horizontalPodAutoscalerSpec.libsonnet'), + horizontalPodAutoscalerStatus: (import 'horizontalPodAutoscalerStatus.libsonnet'), + scale: (import 'scale.libsonnet'), + scaleSpec: (import 'scaleSpec.libsonnet'), + scaleStatus: (import 'scaleStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scale.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scale.libsonnet new file mode 100644 index 00000000000..b0fc338a261 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scale.libsonnet @@ -0,0 +1,61 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scale', url='', help='Scale represents a scaling request for a resource.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Scale', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'autoscaling/v1', + kind: 'Scale', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='ScaleSpec describes the attributes of a scale subresource.'), + spec: { + '#withReplicas':: d.fn(help='desired number of instances for the scaled object.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleSpec.libsonnet new file mode 100644 index 00000000000..ce55d2e4c94 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleSpec.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scaleSpec', url='', help='ScaleSpec describes the attributes of a scale subresource.'), + '#withReplicas':: d.fn(help='desired number of instances for the scaled object.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleStatus.libsonnet new file mode 100644 index 00000000000..bf3063385f7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v1/scaleStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scaleStatus', url='', help='ScaleStatus represents the current status of a scale subresource.'), + '#withReplicas':: d.fn(help='actual number of observed instances of the scaled object.', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withSelector':: d.fn(help='label query over pods that should match the replicas count. This is same as the label selector but in the string format to avoid introspection by clients. The string will be in the same format as the query-param syntax. More info about label selectors: http://kubernetes.io/docs/user-guide/labels#label-selectors', args=[d.arg(name='selector', type=d.T.string)]), + withSelector(selector): { selector: selector }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/crossVersionObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/crossVersionObjectReference.libsonnet new file mode 100644 index 00000000000..b5eb31cc2e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/crossVersionObjectReference.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='crossVersionObjectReference', url='', help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricSource.libsonnet new file mode 100644 index 00000000000..8551ffcb4f5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricSource.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='externalMetricSource', url='', help='ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). Exactly one "target" type should be set.'), + '#metricSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + metricSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metricSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metricSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metricSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metricSelector+: { matchLabels+: matchLabels } }, + }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { targetAverageValue: targetAverageValue }, + '#withTargetValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetValue', type=d.T.string)]), + withTargetValue(targetValue): { targetValue: targetValue }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricStatus.libsonnet new file mode 100644 index 00000000000..3bee0eb3c0f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/externalMetricStatus.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='externalMetricStatus', url='', help='ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object.'), + '#metricSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + metricSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metricSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metricSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metricSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metricSelector+: { matchLabels+: matchLabels } }, + }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { currentAverageValue: currentAverageValue }, + '#withCurrentValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentValue', type=d.T.string)]), + withCurrentValue(currentValue): { currentValue: currentValue }, + '#withMetricName':: d.fn(help='metricName is the name of a metric used for autoscaling in metric system.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscaler.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscaler.libsonnet new file mode 100644 index 00000000000..5f5ecc1b32e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscaler.libsonnet @@ -0,0 +1,76 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscaler', url='', help='HorizontalPodAutoscaler is the configuration for a horizontal pod autoscaler, which automatically manages the replica count of any resource implementing the scale subresource based on the metrics specified.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Horizontalpodautoscaler', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'autoscaling/v2beta1', + kind: 'HorizontalPodAutoscaler', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler.'), + spec: { + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { scaleTargetRef+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { scaleTargetRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { scaleTargetRef+: { name: name } } }, + }, + '#withMaxReplicas':: d.fn(help='maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { spec+: { maxReplicas: maxReplicas } }, + '#withMetrics':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond.', args=[d.arg(name='metrics', type=d.T.array)]), + withMetrics(metrics): { spec+: { metrics: if std.isArray(v=metrics) then metrics else [metrics] } }, + '#withMetricsMixin':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metrics', type=d.T.array)]), + withMetricsMixin(metrics): { spec+: { metrics+: if std.isArray(v=metrics) then metrics else [metrics] } }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { spec+: { minReplicas: minReplicas } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerCondition.libsonnet new file mode 100644 index 00000000000..257ee9862b4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerCondition', url='', help='HorizontalPodAutoscalerCondition describes the state of a HorizontalPodAutoscaler at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='message is a human-readable explanation containing details about the transition', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="reason is the reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='type describes the current condition', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerSpec.libsonnet new file mode 100644 index 00000000000..7b1b7a6e468 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerSpec.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerSpec', url='', help='HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler.'), + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { scaleTargetRef+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { scaleTargetRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleTargetRef+: { name: name } }, + }, + '#withMaxReplicas':: d.fn(help='maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { maxReplicas: maxReplicas }, + '#withMetrics':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond.', args=[d.arg(name='metrics', type=d.T.array)]), + withMetrics(metrics): { metrics: if std.isArray(v=metrics) then metrics else [metrics] }, + '#withMetricsMixin':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metrics', type=d.T.array)]), + withMetricsMixin(metrics): { metrics+: if std.isArray(v=metrics) then metrics else [metrics] }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { minReplicas: minReplicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerStatus.libsonnet new file mode 100644 index 00000000000..93af309bcc3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/horizontalPodAutoscalerStatus.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerStatus', url='', help='HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler.'), + '#withConditions':: d.fn(help='conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withCurrentMetrics':: d.fn(help='currentMetrics is the last read state of the metrics used by this autoscaler.', args=[d.arg(name='currentMetrics', type=d.T.array)]), + withCurrentMetrics(currentMetrics): { currentMetrics: if std.isArray(v=currentMetrics) then currentMetrics else [currentMetrics] }, + '#withCurrentMetricsMixin':: d.fn(help='currentMetrics is the last read state of the metrics used by this autoscaler.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='currentMetrics', type=d.T.array)]), + withCurrentMetricsMixin(currentMetrics): { currentMetrics+: if std.isArray(v=currentMetrics) then currentMetrics else [currentMetrics] }, + '#withCurrentReplicas':: d.fn(help='currentReplicas is current number of replicas of pods managed by this autoscaler, as last seen by the autoscaler.', args=[d.arg(name='currentReplicas', type=d.T.integer)]), + withCurrentReplicas(currentReplicas): { currentReplicas: currentReplicas }, + '#withDesiredReplicas':: d.fn(help='desiredReplicas is the desired number of replicas of pods managed by this autoscaler, as last calculated by the autoscaler.', args=[d.arg(name='desiredReplicas', type=d.T.integer)]), + withDesiredReplicas(desiredReplicas): { desiredReplicas: desiredReplicas }, + '#withLastScaleTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastScaleTime', type=d.T.string)]), + withLastScaleTime(lastScaleTime): { lastScaleTime: lastScaleTime }, + '#withObservedGeneration':: d.fn(help='observedGeneration is the most recent generation observed by this autoscaler.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/main.libsonnet new file mode 100644 index 00000000000..4f75936de0c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/main.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v2beta1', url='', help=''), + crossVersionObjectReference: (import 'crossVersionObjectReference.libsonnet'), + externalMetricSource: (import 'externalMetricSource.libsonnet'), + externalMetricStatus: (import 'externalMetricStatus.libsonnet'), + horizontalPodAutoscaler: (import 'horizontalPodAutoscaler.libsonnet'), + horizontalPodAutoscalerCondition: (import 'horizontalPodAutoscalerCondition.libsonnet'), + horizontalPodAutoscalerSpec: (import 'horizontalPodAutoscalerSpec.libsonnet'), + horizontalPodAutoscalerStatus: (import 'horizontalPodAutoscalerStatus.libsonnet'), + metricSpec: (import 'metricSpec.libsonnet'), + metricStatus: (import 'metricStatus.libsonnet'), + objectMetricSource: (import 'objectMetricSource.libsonnet'), + objectMetricStatus: (import 'objectMetricStatus.libsonnet'), + podsMetricSource: (import 'podsMetricSource.libsonnet'), + podsMetricStatus: (import 'podsMetricStatus.libsonnet'), + resourceMetricSource: (import 'resourceMetricSource.libsonnet'), + resourceMetricStatus: (import 'resourceMetricStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricSpec.libsonnet new file mode 100644 index 00000000000..67a0b79db19 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricSpec.libsonnet @@ -0,0 +1,84 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricSpec', url='', help='MetricSpec specifies how to scale based on a single metric (only `type` and one other matching field should be set at once).'), + '#external':: d.obj(help='ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). Exactly one "target" type should be set.'), + external: { + '#metricSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + metricSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { external+: { metricSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { external+: { metricSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { external+: { metricSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { external+: { metricSelector+: { matchLabels+: matchLabels } } }, + }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { external+: { metricName: metricName } }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { external+: { targetAverageValue: targetAverageValue } }, + '#withTargetValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetValue', type=d.T.string)]), + withTargetValue(targetValue): { external+: { targetValue: targetValue } }, + }, + '#object':: d.obj(help='ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + object: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { object+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { object+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { object+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { object+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#target':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + target: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { object+: { target+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { object+: { target+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { target+: { name: name } } }, + }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { object+: { averageValue: averageValue } }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { object+: { metricName: metricName } }, + '#withTargetValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetValue', type=d.T.string)]), + withTargetValue(targetValue): { object+: { targetValue: targetValue } }, + }, + '#pods':: d.obj(help='PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value.'), + pods: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { pods+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { pods+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { pods+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { pods+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { pods+: { metricName: metricName } }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { pods+: { targetAverageValue: targetAverageValue } }, + }, + '#resource':: d.obj(help='ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set.'), + resource: { + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + '#withTargetAverageUtilization':: d.fn(help='targetAverageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='targetAverageUtilization', type=d.T.integer)]), + withTargetAverageUtilization(targetAverageUtilization): { resource+: { targetAverageUtilization: targetAverageUtilization } }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { resource+: { targetAverageValue: targetAverageValue } }, + }, + '#withType':: d.fn(help='type is the type of metric source. It should be one of "Object", "Pods" or "Resource", each mapping to a matching field in the object.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricStatus.libsonnet new file mode 100644 index 00000000000..d10bd0624f0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/metricStatus.libsonnet @@ -0,0 +1,84 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricStatus', url='', help='MetricStatus describes the last-read state of a single metric.'), + '#external':: d.obj(help='ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object.'), + external: { + '#metricSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + metricSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { external+: { metricSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { external+: { metricSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { external+: { metricSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { external+: { metricSelector+: { matchLabels+: matchLabels } } }, + }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { external+: { currentAverageValue: currentAverageValue } }, + '#withCurrentValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentValue', type=d.T.string)]), + withCurrentValue(currentValue): { external+: { currentValue: currentValue } }, + '#withMetricName':: d.fn(help='metricName is the name of a metric used for autoscaling in metric system.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { external+: { metricName: metricName } }, + }, + '#object':: d.obj(help='ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + object: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { object+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { object+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { object+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { object+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#target':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + target: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { object+: { target+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { object+: { target+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { target+: { name: name } } }, + }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { object+: { averageValue: averageValue } }, + '#withCurrentValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentValue', type=d.T.string)]), + withCurrentValue(currentValue): { object+: { currentValue: currentValue } }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { object+: { metricName: metricName } }, + }, + '#pods':: d.obj(help='PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second).'), + pods: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { pods+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { pods+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { pods+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { pods+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { pods+: { currentAverageValue: currentAverageValue } }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { pods+: { metricName: metricName } }, + }, + '#resource':: d.obj(help='ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source.'), + resource: { + '#withCurrentAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. It will only be present if `targetAverageValue` was set in the corresponding metric specification.', args=[d.arg(name='currentAverageUtilization', type=d.T.integer)]), + withCurrentAverageUtilization(currentAverageUtilization): { resource+: { currentAverageUtilization: currentAverageUtilization } }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { resource+: { currentAverageValue: currentAverageValue } }, + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#withType':: d.fn(help='type is the type of metric source. It will be one of "Object", "Pods" or "Resource", each corresponds to a matching field in the object.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricSource.libsonnet new file mode 100644 index 00000000000..f3448c1b4bd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricSource.libsonnet @@ -0,0 +1,32 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectMetricSource', url='', help='ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#target':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + target: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { target+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { target+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { target+: { name: name } }, + }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { averageValue: averageValue }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#withTargetValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetValue', type=d.T.string)]), + withTargetValue(targetValue): { targetValue: targetValue }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricStatus.libsonnet new file mode 100644 index 00000000000..94b62be174c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/objectMetricStatus.libsonnet @@ -0,0 +1,32 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectMetricStatus', url='', help='ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#target':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + target: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { target+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { target+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { target+: { name: name } }, + }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { averageValue: averageValue }, + '#withCurrentValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentValue', type=d.T.string)]), + withCurrentValue(currentValue): { currentValue: currentValue }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question.', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricSource.libsonnet new file mode 100644 index 00000000000..92085c51331 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricSource.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podsMetricSource', url='', help='PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { targetAverageValue: targetAverageValue }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricStatus.libsonnet new file mode 100644 index 00000000000..d33f4b573d0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/podsMetricStatus.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podsMetricStatus', url='', help='PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second).'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { currentAverageValue: currentAverageValue }, + '#withMetricName':: d.fn(help='metricName is the name of the metric in question', args=[d.arg(name='metricName', type=d.T.string)]), + withMetricName(metricName): { metricName: metricName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricSource.libsonnet new file mode 100644 index 00000000000..f3d0dc447e3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceMetricSource', url='', help='ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set.'), + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withTargetAverageUtilization':: d.fn(help='targetAverageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='targetAverageUtilization', type=d.T.integer)]), + withTargetAverageUtilization(targetAverageUtilization): { targetAverageUtilization: targetAverageUtilization }, + '#withTargetAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='targetAverageValue', type=d.T.string)]), + withTargetAverageValue(targetAverageValue): { targetAverageValue: targetAverageValue }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricStatus.libsonnet new file mode 100644 index 00000000000..5a1c5ec3863 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta1/resourceMetricStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceMetricStatus', url='', help='ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source.'), + '#withCurrentAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. It will only be present if `targetAverageValue` was set in the corresponding metric specification.', args=[d.arg(name='currentAverageUtilization', type=d.T.integer)]), + withCurrentAverageUtilization(currentAverageUtilization): { currentAverageUtilization: currentAverageUtilization }, + '#withCurrentAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='currentAverageValue', type=d.T.string)]), + withCurrentAverageValue(currentAverageValue): { currentAverageValue: currentAverageValue }, + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/crossVersionObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/crossVersionObjectReference.libsonnet new file mode 100644 index 00000000000..b5eb31cc2e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/crossVersionObjectReference.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='crossVersionObjectReference', url='', help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricSource.libsonnet new file mode 100644 index 00000000000..6acc212e119 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricSource.libsonnet @@ -0,0 +1,33 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='externalMetricSource', url='', help='ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster).'), + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { target+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { target+: { averageValue: averageValue } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { target+: { type: type } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { target+: { value: value } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricStatus.libsonnet new file mode 100644 index 00000000000..853b230e2a4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/externalMetricStatus.libsonnet @@ -0,0 +1,31 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='externalMetricStatus', url='', help='ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object.'), + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { current+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { current+: { averageValue: averageValue } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { current+: { value: value } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscaler.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscaler.libsonnet new file mode 100644 index 00000000000..f4c7fa78b09 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscaler.libsonnet @@ -0,0 +1,101 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscaler', url='', help='HorizontalPodAutoscaler is the configuration for a horizontal pod autoscaler, which automatically manages the replica count of any resource implementing the scale subresource based on the metrics specified.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Horizontalpodautoscaler', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'autoscaling/v2beta2', + kind: 'HorizontalPodAutoscaler', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler.'), + spec: { + '#behavior':: d.obj(help='HorizontalPodAutoscalerBehavior configures the scaling behavior of the target in both Up and Down directions (scaleUp and scaleDown fields respectively).'), + behavior: { + '#scaleDown':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleDown: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { spec+: { behavior+: { scaleDown+: { policies: if std.isArray(v=policies) then policies else [policies] } } } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { spec+: { behavior+: { scaleDown+: { policies+: if std.isArray(v=policies) then policies else [policies] } } } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { spec+: { behavior+: { scaleDown+: { selectPolicy: selectPolicy } } } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { spec+: { behavior+: { scaleDown+: { stabilizationWindowSeconds: stabilizationWindowSeconds } } } }, + }, + '#scaleUp':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleUp: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { spec+: { behavior+: { scaleUp+: { policies: if std.isArray(v=policies) then policies else [policies] } } } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { spec+: { behavior+: { scaleUp+: { policies+: if std.isArray(v=policies) then policies else [policies] } } } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { spec+: { behavior+: { scaleUp+: { selectPolicy: selectPolicy } } } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { spec+: { behavior+: { scaleUp+: { stabilizationWindowSeconds: stabilizationWindowSeconds } } } }, + }, + }, + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { scaleTargetRef+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { scaleTargetRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { scaleTargetRef+: { name: name } } }, + }, + '#withMaxReplicas':: d.fn(help='maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { spec+: { maxReplicas: maxReplicas } }, + '#withMetrics':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization.', args=[d.arg(name='metrics', type=d.T.array)]), + withMetrics(metrics): { spec+: { metrics: if std.isArray(v=metrics) then metrics else [metrics] } }, + '#withMetricsMixin':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metrics', type=d.T.array)]), + withMetricsMixin(metrics): { spec+: { metrics+: if std.isArray(v=metrics) then metrics else [metrics] } }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { spec+: { minReplicas: minReplicas } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerBehavior.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerBehavior.libsonnet new file mode 100644 index 00000000000..74248296722 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerBehavior.libsonnet @@ -0,0 +1,28 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerBehavior', url='', help='HorizontalPodAutoscalerBehavior configures the scaling behavior of the target in both Up and Down directions (scaleUp and scaleDown fields respectively).'), + '#scaleDown':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleDown: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { scaleDown+: { policies: if std.isArray(v=policies) then policies else [policies] } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { scaleDown+: { policies+: if std.isArray(v=policies) then policies else [policies] } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { scaleDown+: { selectPolicy: selectPolicy } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { scaleDown+: { stabilizationWindowSeconds: stabilizationWindowSeconds } }, + }, + '#scaleUp':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleUp: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { scaleUp+: { policies: if std.isArray(v=policies) then policies else [policies] } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { scaleUp+: { policies+: if std.isArray(v=policies) then policies else [policies] } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { scaleUp+: { selectPolicy: selectPolicy } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { scaleUp+: { stabilizationWindowSeconds: stabilizationWindowSeconds } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerCondition.libsonnet new file mode 100644 index 00000000000..257ee9862b4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerCondition', url='', help='HorizontalPodAutoscalerCondition describes the state of a HorizontalPodAutoscaler at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='message is a human-readable explanation containing details about the transition', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="reason is the reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='type describes the current condition', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerSpec.libsonnet new file mode 100644 index 00000000000..33e95a22cc5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerSpec.libsonnet @@ -0,0 +1,48 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerSpec', url='', help='HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler.'), + '#behavior':: d.obj(help='HorizontalPodAutoscalerBehavior configures the scaling behavior of the target in both Up and Down directions (scaleUp and scaleDown fields respectively).'), + behavior: { + '#scaleDown':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleDown: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { behavior+: { scaleDown+: { policies: if std.isArray(v=policies) then policies else [policies] } } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { behavior+: { scaleDown+: { policies+: if std.isArray(v=policies) then policies else [policies] } } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { behavior+: { scaleDown+: { selectPolicy: selectPolicy } } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { behavior+: { scaleDown+: { stabilizationWindowSeconds: stabilizationWindowSeconds } } }, + }, + '#scaleUp':: d.obj(help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + scaleUp: { + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { behavior+: { scaleUp+: { policies: if std.isArray(v=policies) then policies else [policies] } } }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { behavior+: { scaleUp+: { policies+: if std.isArray(v=policies) then policies else [policies] } } }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { behavior+: { scaleUp+: { selectPolicy: selectPolicy } } }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { behavior+: { scaleUp+: { stabilizationWindowSeconds: stabilizationWindowSeconds } } }, + }, + }, + '#scaleTargetRef':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + scaleTargetRef: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { scaleTargetRef+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { scaleTargetRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleTargetRef+: { name: name } }, + }, + '#withMaxReplicas':: d.fn(help='maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas.', args=[d.arg(name='maxReplicas', type=d.T.integer)]), + withMaxReplicas(maxReplicas): { maxReplicas: maxReplicas }, + '#withMetrics':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization.', args=[d.arg(name='metrics', type=d.T.array)]), + withMetrics(metrics): { metrics: if std.isArray(v=metrics) then metrics else [metrics] }, + '#withMetricsMixin':: d.fn(help='metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metrics', type=d.T.array)]), + withMetricsMixin(metrics): { metrics+: if std.isArray(v=metrics) then metrics else [metrics] }, + '#withMinReplicas':: d.fn(help='minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. minReplicas is allowed to be 0 if the alpha feature gate HPAScaleToZero is enabled and at least one Object or External metric is configured. Scaling is active as long as at least one metric value is available.', args=[d.arg(name='minReplicas', type=d.T.integer)]), + withMinReplicas(minReplicas): { minReplicas: minReplicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerStatus.libsonnet new file mode 100644 index 00000000000..93af309bcc3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/horizontalPodAutoscalerStatus.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='horizontalPodAutoscalerStatus', url='', help='HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler.'), + '#withConditions':: d.fn(help='conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withCurrentMetrics':: d.fn(help='currentMetrics is the last read state of the metrics used by this autoscaler.', args=[d.arg(name='currentMetrics', type=d.T.array)]), + withCurrentMetrics(currentMetrics): { currentMetrics: if std.isArray(v=currentMetrics) then currentMetrics else [currentMetrics] }, + '#withCurrentMetricsMixin':: d.fn(help='currentMetrics is the last read state of the metrics used by this autoscaler.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='currentMetrics', type=d.T.array)]), + withCurrentMetricsMixin(currentMetrics): { currentMetrics+: if std.isArray(v=currentMetrics) then currentMetrics else [currentMetrics] }, + '#withCurrentReplicas':: d.fn(help='currentReplicas is current number of replicas of pods managed by this autoscaler, as last seen by the autoscaler.', args=[d.arg(name='currentReplicas', type=d.T.integer)]), + withCurrentReplicas(currentReplicas): { currentReplicas: currentReplicas }, + '#withDesiredReplicas':: d.fn(help='desiredReplicas is the desired number of replicas of pods managed by this autoscaler, as last calculated by the autoscaler.', args=[d.arg(name='desiredReplicas', type=d.T.integer)]), + withDesiredReplicas(desiredReplicas): { desiredReplicas: desiredReplicas }, + '#withLastScaleTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastScaleTime', type=d.T.string)]), + withLastScaleTime(lastScaleTime): { lastScaleTime: lastScaleTime }, + '#withObservedGeneration':: d.fn(help='observedGeneration is the most recent generation observed by this autoscaler.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingPolicy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingPolicy.libsonnet new file mode 100644 index 00000000000..2538c7a6ed5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingPolicy.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='hpaScalingPolicy', url='', help='HPAScalingPolicy is a single policy which must hold true for a specified past interval.'), + '#withPeriodSeconds':: d.fn(help='PeriodSeconds specifies the window of time for which the policy should hold true. PeriodSeconds must be greater than zero and less than or equal to 1800 (30 min).', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { periodSeconds: periodSeconds }, + '#withType':: d.fn(help='Type is used to specify the scaling policy.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#withValue':: d.fn(help='Value contains the amount of change which is permitted by the policy. It must be greater than zero', args=[d.arg(name='value', type=d.T.integer)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingRules.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingRules.libsonnet new file mode 100644 index 00000000000..d45a3e289a0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/hpaScalingRules.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='hpaScalingRules', url='', help='HPAScalingRules configures the scaling behavior for one direction. These Rules are applied after calculating DesiredReplicas from metrics for the HPA. They can limit the scaling velocity by specifying scaling policies. They can prevent flapping by specifying the stabilization window, so that the number of replicas is not set instantly, instead, the safest value from the stabilization window is chosen.'), + '#withPolicies':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid', args=[d.arg(name='policies', type=d.T.array)]), + withPolicies(policies): { policies: if std.isArray(v=policies) then policies else [policies] }, + '#withPoliciesMixin':: d.fn(help='policies is a list of potential scaling polices which can be used during scaling. At least one policy must be specified, otherwise the HPAScalingRules will be discarded as invalid\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policies', type=d.T.array)]), + withPoliciesMixin(policies): { policies+: if std.isArray(v=policies) then policies else [policies] }, + '#withSelectPolicy':: d.fn(help='selectPolicy is used to specify which policy should be used. If not set, the default value MaxPolicySelect is used.', args=[d.arg(name='selectPolicy', type=d.T.string)]), + withSelectPolicy(selectPolicy): { selectPolicy: selectPolicy }, + '#withStabilizationWindowSeconds':: d.fn(help='StabilizationWindowSeconds is the number of seconds for which past recommendations should be considered while scaling up or scaling down. StabilizationWindowSeconds must be greater than or equal to zero and less than or equal to 3600 (one hour). If not set, use the default values: - For scale up: 0 (i.e. no stabilization is done). - For scale down: 300 (i.e. the stabilization window is 300 seconds long).', args=[d.arg(name='stabilizationWindowSeconds', type=d.T.integer)]), + withStabilizationWindowSeconds(stabilizationWindowSeconds): { stabilizationWindowSeconds: stabilizationWindowSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/main.libsonnet new file mode 100644 index 00000000000..4ae2fe36034 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/main.libsonnet @@ -0,0 +1,25 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v2beta2', url='', help=''), + crossVersionObjectReference: (import 'crossVersionObjectReference.libsonnet'), + externalMetricSource: (import 'externalMetricSource.libsonnet'), + externalMetricStatus: (import 'externalMetricStatus.libsonnet'), + horizontalPodAutoscaler: (import 'horizontalPodAutoscaler.libsonnet'), + horizontalPodAutoscalerBehavior: (import 'horizontalPodAutoscalerBehavior.libsonnet'), + horizontalPodAutoscalerCondition: (import 'horizontalPodAutoscalerCondition.libsonnet'), + horizontalPodAutoscalerSpec: (import 'horizontalPodAutoscalerSpec.libsonnet'), + horizontalPodAutoscalerStatus: (import 'horizontalPodAutoscalerStatus.libsonnet'), + hpaScalingPolicy: (import 'hpaScalingPolicy.libsonnet'), + hpaScalingRules: (import 'hpaScalingRules.libsonnet'), + metricIdentifier: (import 'metricIdentifier.libsonnet'), + metricSpec: (import 'metricSpec.libsonnet'), + metricStatus: (import 'metricStatus.libsonnet'), + metricTarget: (import 'metricTarget.libsonnet'), + metricValueStatus: (import 'metricValueStatus.libsonnet'), + objectMetricSource: (import 'objectMetricSource.libsonnet'), + objectMetricStatus: (import 'objectMetricStatus.libsonnet'), + podsMetricSource: (import 'podsMetricSource.libsonnet'), + podsMetricStatus: (import 'podsMetricStatus.libsonnet'), + resourceMetricSource: (import 'resourceMetricSource.libsonnet'), + resourceMetricStatus: (import 'resourceMetricStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricIdentifier.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricIdentifier.libsonnet new file mode 100644 index 00000000000..31bebd295f0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricIdentifier.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricIdentifier', url='', help='MetricIdentifier defines the name and optionally selector for a metric'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricSpec.libsonnet new file mode 100644 index 00000000000..d49dfd0bf32 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricSpec.libsonnet @@ -0,0 +1,123 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricSpec', url='', help='MetricSpec specifies how to scale based on a single metric (only `type` and one other matching field should be set at once).'), + '#external':: d.obj(help='ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster).'), + external: { + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { external+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { external+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { external+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { external+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { external+: { metric+: { name: name } } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { external+: { target+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { external+: { target+: { averageValue: averageValue } } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { external+: { target+: { type: type } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { external+: { target+: { value: value } } }, + }, + }, + '#object':: d.obj(help='ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + object: { + '#describedObject':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + describedObject: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { object+: { describedObject+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { object+: { describedObject+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { describedObject+: { name: name } } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { object+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { object+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { object+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { object+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { metric+: { name: name } } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { object+: { target+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { object+: { target+: { averageValue: averageValue } } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { object+: { target+: { type: type } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { object+: { target+: { value: value } } }, + }, + }, + '#pods':: d.obj(help='PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value.'), + pods: { + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { pods+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { pods+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { pods+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { pods+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { pods+: { metric+: { name: name } } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { pods+: { target+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { pods+: { target+: { averageValue: averageValue } } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { pods+: { target+: { type: type } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { pods+: { target+: { value: value } } }, + }, + }, + '#resource':: d.obj(help='ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set.'), + resource: { + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { resource+: { target+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { resource+: { target+: { averageValue: averageValue } } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { resource+: { target+: { type: type } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { resource+: { target+: { value: value } } }, + }, + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#withType':: d.fn(help='type is the type of metric source. It should be one of "Object", "Pods" or "Resource", each mapping to a matching field in the object.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricStatus.libsonnet new file mode 100644 index 00000000000..cfa54e7d1d3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricStatus.libsonnet @@ -0,0 +1,115 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricStatus', url='', help='MetricStatus describes the last-read state of a single metric.'), + '#external':: d.obj(help='ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object.'), + external: { + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { external+: { current+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { external+: { current+: { averageValue: averageValue } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { external+: { current+: { value: value } } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { external+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { external+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { external+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { external+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { external+: { metric+: { name: name } } }, + }, + }, + '#object':: d.obj(help='ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + object: { + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { object+: { current+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { object+: { current+: { averageValue: averageValue } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { object+: { current+: { value: value } } }, + }, + '#describedObject':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + describedObject: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { object+: { describedObject+: { apiVersion: apiVersion } } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { object+: { describedObject+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { describedObject+: { name: name } } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { object+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { object+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { object+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { object+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { object+: { metric+: { name: name } } }, + }, + }, + '#pods':: d.obj(help='PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second).'), + pods: { + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { pods+: { current+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { pods+: { current+: { averageValue: averageValue } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { pods+: { current+: { value: value } } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { pods+: { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { pods+: { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { pods+: { metric+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { pods+: { metric+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { pods+: { metric+: { name: name } } }, + }, + }, + '#resource':: d.obj(help='ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source.'), + resource: { + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { resource+: { current+: { averageUtilization: averageUtilization } } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { resource+: { current+: { averageValue: averageValue } } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { resource+: { current+: { value: value } } }, + }, + '#withName':: d.fn(help='Name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#withType':: d.fn(help='type is the type of metric source. It will be one of "Object", "Pods" or "Resource", each corresponds to a matching field in the object.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricTarget.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricTarget.libsonnet new file mode 100644 index 00000000000..9e0add77400 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricTarget.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricTarget', url='', help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { averageUtilization: averageUtilization }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { averageValue: averageValue }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricValueStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricValueStatus.libsonnet new file mode 100644 index 00000000000..35e9e68492f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/metricValueStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='metricValueStatus', url='', help='MetricValueStatus holds the current value for a metric'), + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { averageUtilization: averageUtilization }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { averageValue: averageValue }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricSource.libsonnet new file mode 100644 index 00000000000..f360b97e84c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricSource.libsonnet @@ -0,0 +1,42 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectMetricSource', url='', help='ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + '#describedObject':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + describedObject: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { describedObject+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { describedObject+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { describedObject+: { name: name } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { target+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { target+: { averageValue: averageValue } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { target+: { type: type } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { target+: { value: value } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricStatus.libsonnet new file mode 100644 index 00000000000..ea885293c17 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/objectMetricStatus.libsonnet @@ -0,0 +1,40 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectMetricStatus', url='', help='ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object).'), + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { current+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { current+: { averageValue: averageValue } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { current+: { value: value } }, + }, + '#describedObject':: d.obj(help='CrossVersionObjectReference contains enough information to let you identify the referred resource.'), + describedObject: { + '#withApiVersion':: d.fn(help='API version of the referent', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { describedObject+: { apiVersion: apiVersion } }, + '#withKind':: d.fn(help='Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { describedObject+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { describedObject+: { name: name } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricSource.libsonnet new file mode 100644 index 00000000000..cba73286820 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricSource.libsonnet @@ -0,0 +1,33 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podsMetricSource', url='', help='PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value.'), + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { target+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { target+: { averageValue: averageValue } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { target+: { type: type } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { target+: { value: value } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricStatus.libsonnet new file mode 100644 index 00000000000..67deaad7dd5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/podsMetricStatus.libsonnet @@ -0,0 +1,31 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podsMetricStatus', url='', help='PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second).'), + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { current+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { current+: { averageValue: averageValue } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { current+: { value: value } }, + }, + '#metric':: d.obj(help='MetricIdentifier defines the name and optionally selector for a metric'), + metric: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { metric+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { metric+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { metric+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { metric+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withName':: d.fn(help='name is the name of the given metric', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metric+: { name: name } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricSource.libsonnet new file mode 100644 index 00000000000..de5248cba81 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricSource.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceMetricSource', url='', help='ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set.'), + '#target':: d.obj(help='MetricTarget defines the target value, average value, or average utilization of a specific metric'), + target: { + '#withAverageUtilization':: d.fn(help='averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { target+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { target+: { averageValue: averageValue } }, + '#withType':: d.fn(help='type represents whether the metric type is Utilization, Value, or AverageValue', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { target+: { type: type } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { target+: { value: value } }, + }, + '#withName':: d.fn(help='name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricStatus.libsonnet new file mode 100644 index 00000000000..5e4623bb886 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/autoscaling/v2beta2/resourceMetricStatus.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceMetricStatus', url='', help='ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source.'), + '#current':: d.obj(help='MetricValueStatus holds the current value for a metric'), + current: { + '#withAverageUtilization':: d.fn(help='currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods.', args=[d.arg(name='averageUtilization', type=d.T.integer)]), + withAverageUtilization(averageUtilization): { current+: { averageUtilization: averageUtilization } }, + '#withAverageValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='averageValue', type=d.T.string)]), + withAverageValue(averageValue): { current+: { averageValue: averageValue } }, + '#withValue':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { current+: { value: value } }, + }, + '#withName':: d.fn(help='Name is the name of the resource in question.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/main.libsonnet new file mode 100644 index 00000000000..7c9faf5243b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='batch', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), + v2alpha1: (import 'v2alpha1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/job.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/job.libsonnet new file mode 100644 index 00000000000..6601290957f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/job.libsonnet @@ -0,0 +1,321 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='job', url='', help='Job represents the configuration of a single job.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Job', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'batch/v1', + kind: 'Job', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { spec+: { backoffLimit: backoffLimit } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { spec+: { completions: completions } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { spec+: { manualSelector: manualSelector } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { spec+: { parallelism: parallelism } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobCondition.libsonnet new file mode 100644 index 00000000000..4fddec451bc --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='jobCondition', url='', help='JobCondition describes current state of a job.'), + '#withLastProbeTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastProbeTime', type=d.T.string)]), + withLastProbeTime(lastProbeTime): { lastProbeTime: lastProbeTime }, + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="(brief) reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of job condition, Complete or Failed.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobSpec.libsonnet new file mode 100644 index 00000000000..391e7349087 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobSpec.libsonnet @@ -0,0 +1,268 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='jobSpec', url='', help='JobSpec describes how the job execution will look like.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { activeDeadlineSeconds: activeDeadlineSeconds }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { backoffLimit: backoffLimit }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { completions: completions }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { manualSelector: manualSelector }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { parallelism: parallelism }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { ttlSecondsAfterFinished: ttlSecondsAfterFinished }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobStatus.libsonnet new file mode 100644 index 00000000000..7735af12c32 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/jobStatus.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='jobStatus', url='', help='JobStatus represents the current state of a Job.'), + '#withActive':: d.fn(help='The number of actively running pods.', args=[d.arg(name='active', type=d.T.integer)]), + withActive(active): { active: active }, + '#withCompletionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='completionTime', type=d.T.string)]), + withCompletionTime(completionTime): { completionTime: completionTime }, + '#withConditions':: d.fn(help="The latest available observations of an object's current state. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="The latest available observations of an object's current state. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withFailed':: d.fn(help='The number of pods which reached phase Failed.', args=[d.arg(name='failed', type=d.T.integer)]), + withFailed(failed): { failed: failed }, + '#withStartTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startTime', type=d.T.string)]), + withStartTime(startTime): { startTime: startTime }, + '#withSucceeded':: d.fn(help='The number of pods which reached phase Succeeded.', args=[d.arg(name='succeeded', type=d.T.integer)]), + withSucceeded(succeeded): { succeeded: succeeded }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/main.libsonnet new file mode 100644 index 00000000000..0200f1ef26f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + job: (import 'job.libsonnet'), + jobCondition: (import 'jobCondition.libsonnet'), + jobSpec: (import 'jobSpec.libsonnet'), + jobStatus: (import 'jobStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJob.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJob.libsonnet new file mode 100644 index 00000000000..148e1455e3e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJob.libsonnet @@ -0,0 +1,384 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJob', url='', help='CronJob represents the configuration of a single cron job.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Cronjob', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'batch/v1beta1', + kind: 'CronJob', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CronJobSpec describes how the job execution will look like and when it will actually run.'), + spec: { + '#jobTemplate':: d.obj(help='JobTemplateSpec describes the data a Job should have when created from a template'), + jobTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { jobTemplate+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { jobTemplate+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { jobTemplate+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { jobTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { jobTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { jobTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { jobTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { jobTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { jobTemplate+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { jobTemplate+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { jobTemplate+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { jobTemplate+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { jobTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { jobTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { jobTemplate+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { jobTemplate+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { jobTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { jobTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { jobTemplate+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { jobTemplate+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { jobTemplate+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { jobTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { jobTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { jobTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { jobTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { annotations: annotations } } } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { annotations+: annotations } } } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { clusterName: clusterName } } } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { generateName: generateName } } } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { generation: generation } } } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { labels: labels } } } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { labels+: labels } } } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { namespace: namespace } } } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { selfLink: selfLink } } } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { uid: uid } } } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostIPC: hostIPC } } } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostPID: hostPID } } } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostname: hostname } } } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeName: nodeName } } } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { overhead: overhead } } } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { overhead+: overhead } } } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { priority: priority } } } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { schedulerName: schedulerName } } } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { subdomain: subdomain } } } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { jobTemplate+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { spec+: { jobTemplate+: { spec+: { backoffLimit: backoffLimit } } } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { spec+: { jobTemplate+: { spec+: { completions: completions } } } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { spec+: { jobTemplate+: { spec+: { manualSelector: manualSelector } } } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { spec+: { jobTemplate+: { spec+: { parallelism: parallelism } } } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { spec+: { jobTemplate+: { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } } } }, + }, + }, + '#withConcurrencyPolicy':: d.fn(help="Specifies how to treat concurrent executions of a Job. Valid values are: - 'Allow' (default): allows CronJobs to run concurrently; - 'Forbid': forbids concurrent runs, skipping next run if previous run hasn't finished yet; - 'Replace': cancels currently running job and replaces it with a new one", args=[d.arg(name='concurrencyPolicy', type=d.T.string)]), + withConcurrencyPolicy(concurrencyPolicy): { spec+: { concurrencyPolicy: concurrencyPolicy } }, + '#withFailedJobsHistoryLimit':: d.fn(help='The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1.', args=[d.arg(name='failedJobsHistoryLimit', type=d.T.integer)]), + withFailedJobsHistoryLimit(failedJobsHistoryLimit): { spec+: { failedJobsHistoryLimit: failedJobsHistoryLimit } }, + '#withSchedule':: d.fn(help='The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron.', args=[d.arg(name='schedule', type=d.T.string)]), + withSchedule(schedule): { spec+: { schedule: schedule } }, + '#withStartingDeadlineSeconds':: d.fn(help='Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones.', args=[d.arg(name='startingDeadlineSeconds', type=d.T.integer)]), + withStartingDeadlineSeconds(startingDeadlineSeconds): { spec+: { startingDeadlineSeconds: startingDeadlineSeconds } }, + '#withSuccessfulJobsHistoryLimit':: d.fn(help='The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 3.', args=[d.arg(name='successfulJobsHistoryLimit', type=d.T.integer)]), + withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit): { spec+: { successfulJobsHistoryLimit: successfulJobsHistoryLimit } }, + '#withSuspend':: d.fn(help='This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false.', args=[d.arg(name='suspend', type=d.T.boolean)]), + withSuspend(suspend): { spec+: { suspend: suspend } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobSpec.libsonnet new file mode 100644 index 00000000000..cfe60215acd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobSpec.libsonnet @@ -0,0 +1,331 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJobSpec', url='', help='CronJobSpec describes how the job execution will look like and when it will actually run.'), + '#jobTemplate':: d.obj(help='JobTemplateSpec describes the data a Job should have when created from a template'), + jobTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { jobTemplate+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { jobTemplate+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { jobTemplate+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { jobTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { jobTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { jobTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { jobTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { jobTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { jobTemplate+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { jobTemplate+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { jobTemplate+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { jobTemplate+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { jobTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { jobTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { jobTemplate+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { jobTemplate+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { jobTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { jobTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { jobTemplate+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { jobTemplate+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { jobTemplate+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { jobTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { jobTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { jobTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { jobTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { jobTemplate+: { spec+: { template+: { metadata+: { annotations: annotations } } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { jobTemplate+: { spec+: { template+: { metadata+: { annotations+: annotations } } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { jobTemplate+: { spec+: { template+: { metadata+: { clusterName: clusterName } } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { jobTemplate+: { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { jobTemplate+: { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { jobTemplate+: { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { jobTemplate+: { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { jobTemplate+: { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { jobTemplate+: { spec+: { template+: { metadata+: { generateName: generateName } } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { jobTemplate+: { spec+: { template+: { metadata+: { generation: generation } } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { jobTemplate+: { spec+: { template+: { metadata+: { labels: labels } } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { jobTemplate+: { spec+: { template+: { metadata+: { labels+: labels } } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { jobTemplate+: { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { jobTemplate+: { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { jobTemplate+: { spec+: { template+: { metadata+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { jobTemplate+: { spec+: { template+: { metadata+: { namespace: namespace } } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { jobTemplate+: { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { jobTemplate+: { spec+: { template+: { metadata+: { selfLink: selfLink } } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { jobTemplate+: { spec+: { template+: { metadata+: { uid: uid } } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { jobTemplate+: { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { jobTemplate+: { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { jobTemplate+: { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { jobTemplate+: { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { jobTemplate+: { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { jobTemplate+: { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { jobTemplate+: { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { jobTemplate+: { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { jobTemplate+: { spec+: { template+: { spec+: { hostIPC: hostIPC } } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { jobTemplate+: { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { jobTemplate+: { spec+: { template+: { spec+: { hostPID: hostPID } } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { jobTemplate+: { spec+: { template+: { spec+: { hostname: hostname } } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { jobTemplate+: { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { jobTemplate+: { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { jobTemplate+: { spec+: { template+: { spec+: { nodeName: nodeName } } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { jobTemplate+: { spec+: { template+: { spec+: { overhead: overhead } } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { jobTemplate+: { spec+: { template+: { spec+: { overhead+: overhead } } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { jobTemplate+: { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { jobTemplate+: { spec+: { template+: { spec+: { priority: priority } } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { jobTemplate+: { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { jobTemplate+: { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { jobTemplate+: { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { jobTemplate+: { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { jobTemplate+: { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { jobTemplate+: { spec+: { template+: { spec+: { schedulerName: schedulerName } } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { jobTemplate+: { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { jobTemplate+: { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { jobTemplate+: { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { jobTemplate+: { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { jobTemplate+: { spec+: { template+: { spec+: { subdomain: subdomain } } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { jobTemplate+: { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { jobTemplate+: { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { jobTemplate+: { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { jobTemplate+: { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { jobTemplate+: { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { jobTemplate+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { jobTemplate+: { spec+: { backoffLimit: backoffLimit } } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { jobTemplate+: { spec+: { completions: completions } } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { jobTemplate+: { spec+: { manualSelector: manualSelector } } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { jobTemplate+: { spec+: { parallelism: parallelism } } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { jobTemplate+: { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } } }, + }, + }, + '#withConcurrencyPolicy':: d.fn(help="Specifies how to treat concurrent executions of a Job. Valid values are: - 'Allow' (default): allows CronJobs to run concurrently; - 'Forbid': forbids concurrent runs, skipping next run if previous run hasn't finished yet; - 'Replace': cancels currently running job and replaces it with a new one", args=[d.arg(name='concurrencyPolicy', type=d.T.string)]), + withConcurrencyPolicy(concurrencyPolicy): { concurrencyPolicy: concurrencyPolicy }, + '#withFailedJobsHistoryLimit':: d.fn(help='The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1.', args=[d.arg(name='failedJobsHistoryLimit', type=d.T.integer)]), + withFailedJobsHistoryLimit(failedJobsHistoryLimit): { failedJobsHistoryLimit: failedJobsHistoryLimit }, + '#withSchedule':: d.fn(help='The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron.', args=[d.arg(name='schedule', type=d.T.string)]), + withSchedule(schedule): { schedule: schedule }, + '#withStartingDeadlineSeconds':: d.fn(help='Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones.', args=[d.arg(name='startingDeadlineSeconds', type=d.T.integer)]), + withStartingDeadlineSeconds(startingDeadlineSeconds): { startingDeadlineSeconds: startingDeadlineSeconds }, + '#withSuccessfulJobsHistoryLimit':: d.fn(help='The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 3.', args=[d.arg(name='successfulJobsHistoryLimit', type=d.T.integer)]), + withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit): { successfulJobsHistoryLimit: successfulJobsHistoryLimit }, + '#withSuspend':: d.fn(help='This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false.', args=[d.arg(name='suspend', type=d.T.boolean)]), + withSuspend(suspend): { suspend: suspend }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobStatus.libsonnet new file mode 100644 index 00000000000..834dd1932fa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/cronJobStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJobStatus', url='', help='CronJobStatus represents the current state of a cron job.'), + '#withActive':: d.fn(help='A list of pointers to currently running jobs.', args=[d.arg(name='active', type=d.T.array)]), + withActive(active): { active: if std.isArray(v=active) then active else [active] }, + '#withActiveMixin':: d.fn(help='A list of pointers to currently running jobs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='active', type=d.T.array)]), + withActiveMixin(active): { active+: if std.isArray(v=active) then active else [active] }, + '#withLastScheduleTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastScheduleTime', type=d.T.string)]), + withLastScheduleTime(lastScheduleTime): { lastScheduleTime: lastScheduleTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/jobTemplateSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/jobTemplateSpec.libsonnet new file mode 100644 index 00000000000..3cd1c4a84c9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/jobTemplateSpec.libsonnet @@ -0,0 +1,316 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='jobTemplateSpec', url='', help='JobTemplateSpec describes the data a Job should have when created from a template'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { spec+: { backoffLimit: backoffLimit } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { spec+: { completions: completions } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { spec+: { manualSelector: manualSelector } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { spec+: { parallelism: parallelism } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/main.libsonnet new file mode 100644 index 00000000000..e9498f60d44 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v1beta1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + cronJob: (import 'cronJob.libsonnet'), + cronJobSpec: (import 'cronJobSpec.libsonnet'), + cronJobStatus: (import 'cronJobStatus.libsonnet'), + jobTemplateSpec: (import 'jobTemplateSpec.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJob.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJob.libsonnet new file mode 100644 index 00000000000..7020847db74 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJob.libsonnet @@ -0,0 +1,384 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJob', url='', help='CronJob represents the configuration of a single cron job.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Cronjob', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'batch/v2alpha1', + kind: 'CronJob', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CronJobSpec describes how the job execution will look like and when it will actually run.'), + spec: { + '#jobTemplate':: d.obj(help='JobTemplateSpec describes the data a Job should have when created from a template'), + jobTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { jobTemplate+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { jobTemplate+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { jobTemplate+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { jobTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { jobTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { jobTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { jobTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { jobTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { jobTemplate+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { jobTemplate+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { jobTemplate+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { jobTemplate+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { jobTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { jobTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { jobTemplate+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { jobTemplate+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { jobTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { jobTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { jobTemplate+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { jobTemplate+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { jobTemplate+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { jobTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { jobTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { jobTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { jobTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { annotations: annotations } } } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { annotations+: annotations } } } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { clusterName: clusterName } } } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { generateName: generateName } } } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { generation: generation } } } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { labels: labels } } } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { labels+: labels } } } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { namespace: namespace } } } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { selfLink: selfLink } } } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { jobTemplate+: { spec+: { template+: { metadata+: { uid: uid } } } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostIPC: hostIPC } } } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostPID: hostPID } } } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { hostname: hostname } } } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeName: nodeName } } } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { overhead: overhead } } } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { overhead+: overhead } } } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { priority: priority } } } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { schedulerName: schedulerName } } } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { subdomain: subdomain } } } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { jobTemplate+: { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { jobTemplate+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { spec+: { jobTemplate+: { spec+: { backoffLimit: backoffLimit } } } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { spec+: { jobTemplate+: { spec+: { completions: completions } } } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { spec+: { jobTemplate+: { spec+: { manualSelector: manualSelector } } } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { spec+: { jobTemplate+: { spec+: { parallelism: parallelism } } } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { spec+: { jobTemplate+: { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } } } }, + }, + }, + '#withConcurrencyPolicy':: d.fn(help="Specifies how to treat concurrent executions of a Job. Valid values are: - 'Allow' (default): allows CronJobs to run concurrently; - 'Forbid': forbids concurrent runs, skipping next run if previous run hasn't finished yet; - 'Replace': cancels currently running job and replaces it with a new one", args=[d.arg(name='concurrencyPolicy', type=d.T.string)]), + withConcurrencyPolicy(concurrencyPolicy): { spec+: { concurrencyPolicy: concurrencyPolicy } }, + '#withFailedJobsHistoryLimit':: d.fn(help='The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified.', args=[d.arg(name='failedJobsHistoryLimit', type=d.T.integer)]), + withFailedJobsHistoryLimit(failedJobsHistoryLimit): { spec+: { failedJobsHistoryLimit: failedJobsHistoryLimit } }, + '#withSchedule':: d.fn(help='The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron.', args=[d.arg(name='schedule', type=d.T.string)]), + withSchedule(schedule): { spec+: { schedule: schedule } }, + '#withStartingDeadlineSeconds':: d.fn(help='Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones.', args=[d.arg(name='startingDeadlineSeconds', type=d.T.integer)]), + withStartingDeadlineSeconds(startingDeadlineSeconds): { spec+: { startingDeadlineSeconds: startingDeadlineSeconds } }, + '#withSuccessfulJobsHistoryLimit':: d.fn(help='The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified.', args=[d.arg(name='successfulJobsHistoryLimit', type=d.T.integer)]), + withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit): { spec+: { successfulJobsHistoryLimit: successfulJobsHistoryLimit } }, + '#withSuspend':: d.fn(help='This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false.', args=[d.arg(name='suspend', type=d.T.boolean)]), + withSuspend(suspend): { spec+: { suspend: suspend } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobSpec.libsonnet new file mode 100644 index 00000000000..ad96676ab8d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobSpec.libsonnet @@ -0,0 +1,331 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJobSpec', url='', help='CronJobSpec describes how the job execution will look like and when it will actually run.'), + '#jobTemplate':: d.obj(help='JobTemplateSpec describes the data a Job should have when created from a template'), + jobTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { jobTemplate+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { jobTemplate+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { jobTemplate+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { jobTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { jobTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { jobTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { jobTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { jobTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { jobTemplate+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { jobTemplate+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { jobTemplate+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { jobTemplate+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { jobTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { jobTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { jobTemplate+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { jobTemplate+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { jobTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { jobTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { jobTemplate+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { jobTemplate+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { jobTemplate+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { jobTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { jobTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { jobTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { jobTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { jobTemplate+: { spec+: { template+: { metadata+: { annotations: annotations } } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { jobTemplate+: { spec+: { template+: { metadata+: { annotations+: annotations } } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { jobTemplate+: { spec+: { template+: { metadata+: { clusterName: clusterName } } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { jobTemplate+: { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { jobTemplate+: { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { jobTemplate+: { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { jobTemplate+: { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { jobTemplate+: { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { jobTemplate+: { spec+: { template+: { metadata+: { generateName: generateName } } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { jobTemplate+: { spec+: { template+: { metadata+: { generation: generation } } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { jobTemplate+: { spec+: { template+: { metadata+: { labels: labels } } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { jobTemplate+: { spec+: { template+: { metadata+: { labels+: labels } } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { jobTemplate+: { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { jobTemplate+: { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { jobTemplate+: { spec+: { template+: { metadata+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { jobTemplate+: { spec+: { template+: { metadata+: { namespace: namespace } } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { jobTemplate+: { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { jobTemplate+: { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { jobTemplate+: { spec+: { template+: { metadata+: { selfLink: selfLink } } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { jobTemplate+: { spec+: { template+: { metadata+: { uid: uid } } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { jobTemplate+: { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { jobTemplate+: { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { jobTemplate+: { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { jobTemplate+: { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { jobTemplate+: { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { jobTemplate+: { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { jobTemplate+: { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { jobTemplate+: { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { jobTemplate+: { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { jobTemplate+: { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { jobTemplate+: { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { jobTemplate+: { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { jobTemplate+: { spec+: { template+: { spec+: { hostIPC: hostIPC } } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { jobTemplate+: { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { jobTemplate+: { spec+: { template+: { spec+: { hostPID: hostPID } } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { jobTemplate+: { spec+: { template+: { spec+: { hostname: hostname } } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { jobTemplate+: { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { jobTemplate+: { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { jobTemplate+: { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { jobTemplate+: { spec+: { template+: { spec+: { nodeName: nodeName } } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { jobTemplate+: { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { jobTemplate+: { spec+: { template+: { spec+: { overhead: overhead } } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { jobTemplate+: { spec+: { template+: { spec+: { overhead+: overhead } } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { jobTemplate+: { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { jobTemplate+: { spec+: { template+: { spec+: { priority: priority } } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { jobTemplate+: { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { jobTemplate+: { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { jobTemplate+: { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { jobTemplate+: { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { jobTemplate+: { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { jobTemplate+: { spec+: { template+: { spec+: { schedulerName: schedulerName } } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { jobTemplate+: { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { jobTemplate+: { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { jobTemplate+: { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { jobTemplate+: { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { jobTemplate+: { spec+: { template+: { spec+: { subdomain: subdomain } } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { jobTemplate+: { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { jobTemplate+: { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { jobTemplate+: { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { jobTemplate+: { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { jobTemplate+: { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { jobTemplate+: { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { jobTemplate+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { jobTemplate+: { spec+: { backoffLimit: backoffLimit } } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { jobTemplate+: { spec+: { completions: completions } } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { jobTemplate+: { spec+: { manualSelector: manualSelector } } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { jobTemplate+: { spec+: { parallelism: parallelism } } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { jobTemplate+: { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } } }, + }, + }, + '#withConcurrencyPolicy':: d.fn(help="Specifies how to treat concurrent executions of a Job. Valid values are: - 'Allow' (default): allows CronJobs to run concurrently; - 'Forbid': forbids concurrent runs, skipping next run if previous run hasn't finished yet; - 'Replace': cancels currently running job and replaces it with a new one", args=[d.arg(name='concurrencyPolicy', type=d.T.string)]), + withConcurrencyPolicy(concurrencyPolicy): { concurrencyPolicy: concurrencyPolicy }, + '#withFailedJobsHistoryLimit':: d.fn(help='The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified.', args=[d.arg(name='failedJobsHistoryLimit', type=d.T.integer)]), + withFailedJobsHistoryLimit(failedJobsHistoryLimit): { failedJobsHistoryLimit: failedJobsHistoryLimit }, + '#withSchedule':: d.fn(help='The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron.', args=[d.arg(name='schedule', type=d.T.string)]), + withSchedule(schedule): { schedule: schedule }, + '#withStartingDeadlineSeconds':: d.fn(help='Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones.', args=[d.arg(name='startingDeadlineSeconds', type=d.T.integer)]), + withStartingDeadlineSeconds(startingDeadlineSeconds): { startingDeadlineSeconds: startingDeadlineSeconds }, + '#withSuccessfulJobsHistoryLimit':: d.fn(help='The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified.', args=[d.arg(name='successfulJobsHistoryLimit', type=d.T.integer)]), + withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit): { successfulJobsHistoryLimit: successfulJobsHistoryLimit }, + '#withSuspend':: d.fn(help='This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false.', args=[d.arg(name='suspend', type=d.T.boolean)]), + withSuspend(suspend): { suspend: suspend }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobStatus.libsonnet new file mode 100644 index 00000000000..834dd1932fa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/cronJobStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cronJobStatus', url='', help='CronJobStatus represents the current state of a cron job.'), + '#withActive':: d.fn(help='A list of pointers to currently running jobs.', args=[d.arg(name='active', type=d.T.array)]), + withActive(active): { active: if std.isArray(v=active) then active else [active] }, + '#withActiveMixin':: d.fn(help='A list of pointers to currently running jobs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='active', type=d.T.array)]), + withActiveMixin(active): { active+: if std.isArray(v=active) then active else [active] }, + '#withLastScheduleTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastScheduleTime', type=d.T.string)]), + withLastScheduleTime(lastScheduleTime): { lastScheduleTime: lastScheduleTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/jobTemplateSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/jobTemplateSpec.libsonnet new file mode 100644 index 00000000000..3cd1c4a84c9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/jobTemplateSpec.libsonnet @@ -0,0 +1,316 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='jobTemplateSpec', url='', help='JobTemplateSpec describes the data a Job should have when created from a template'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#spec':: d.obj(help='JobSpec describes how the job execution will look like.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } }, + '#withBackoffLimit':: d.fn(help='Specifies the number of retries before marking this job failed. Defaults to 6', args=[d.arg(name='backoffLimit', type=d.T.integer)]), + withBackoffLimit(backoffLimit): { spec+: { backoffLimit: backoffLimit } }, + '#withCompletions':: d.fn(help='Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='completions', type=d.T.integer)]), + withCompletions(completions): { spec+: { completions: completions } }, + '#withManualSelector':: d.fn(help='manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector', args=[d.arg(name='manualSelector', type=d.T.boolean)]), + withManualSelector(manualSelector): { spec+: { manualSelector: manualSelector } }, + '#withParallelism':: d.fn(help='Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/', args=[d.arg(name='parallelism', type=d.T.integer)]), + withParallelism(parallelism): { spec+: { parallelism: parallelism } }, + '#withTtlSecondsAfterFinished':: d.fn(help="ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature.", args=[d.arg(name='ttlSecondsAfterFinished', type=d.T.integer)]), + withTtlSecondsAfterFinished(ttlSecondsAfterFinished): { spec+: { ttlSecondsAfterFinished: ttlSecondsAfterFinished } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/main.libsonnet new file mode 100644 index 00000000000..0393a4c06ea --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/batch/v2alpha1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v2alpha1', url='', help=''), + cronJob: (import 'cronJob.libsonnet'), + cronJobSpec: (import 'cronJobSpec.libsonnet'), + cronJobStatus: (import 'cronJobStatus.libsonnet'), + jobTemplateSpec: (import 'jobTemplateSpec.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/main.libsonnet new file mode 100644 index 00000000000..2fc91e17e6c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificates', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequest.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequest.libsonnet new file mode 100644 index 00000000000..840928e0ad9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequest.libsonnet @@ -0,0 +1,79 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequest', url='', help='CertificateSigningRequest objects provide a mechanism to obtain x509 certificates by submitting a certificate signing request, and having it asynchronously approved and issued.\n\nKubelets use this API to obtain:\n 1. client certificates to authenticate to kube-apiserver (with the "kubernetes.io/kube-apiserver-client-kubelet" signerName).\n 2. serving certificates for TLS endpoints kube-apiserver can connect to securely (with the "kubernetes.io/kubelet-serving" signerName).\n\nThis API can be used to request client certificates to authenticate to kube-apiserver (with the "kubernetes.io/kube-apiserver-client" signerName), or to obtain certificates from custom non-Kubernetes signers.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Certificatesigningrequest', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'certificates.k8s.io/v1', + kind: 'CertificateSigningRequest', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CertificateSigningRequestSpec contains the certificate request.'), + spec: { + '#withExtra':: d.fn(help='extra contains extra attributes of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='extra contains extra attributes of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroups':: d.fn(help='groups contains group membership of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { spec+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help='groups contains group membership of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { spec+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withRequest':: d.fn(help='request contains an x509 certificate signing request encoded in a "CERTIFICATE REQUEST" PEM block. When serialized as JSON or YAML, the data is additionally base64-encoded.', args=[d.arg(name='request', type=d.T.string)]), + withRequest(request): { spec+: { request: request } }, + '#withSignerName':: d.fn(help='signerName indicates the requested signer, and is a qualified name.\n\nList/watch requests for CertificateSigningRequests can filter on this field using a "spec.signerName=NAME" fieldSelector.\n\nWell-known Kubernetes signers are:\n 1. "kubernetes.io/kube-apiserver-client": issues client certificates that can be used to authenticate to kube-apiserver.\n Requests for this signer are never auto-approved by kube-controller-manager, can be issued by the "csrsigning" controller in kube-controller-manager.\n 2. "kubernetes.io/kube-apiserver-client-kubelet": issues client certificates that kubelets use to authenticate to kube-apiserver.\n Requests for this signer can be auto-approved by the "csrapproving" controller in kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager.\n 3. "kubernetes.io/kubelet-serving" issues serving certificates that kubelets use to serve TLS endpoints, which kube-apiserver can connect to securely.\n Requests for this signer are never auto-approved by kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager.\n\nMore details are available at https://k8s.io/docs/reference/access-authn-authz/certificate-signing-requests/#kubernetes-signers\n\nCustom signerNames can also be specified. The signer defines:\n 1. Trust distribution: how trust (CA bundles) are distributed.\n 2. Permitted subjects: and behavior when a disallowed subject is requested.\n 3. Required, permitted, or forbidden x509 extensions in the request (including whether subjectAltNames are allowed, which types, restrictions on allowed values) and behavior when a disallowed extension is requested.\n 4. Required, permitted, or forbidden key usages / extended key usages.\n 5. Expiration/certificate lifetime: whether it is fixed by the signer, configurable by the admin.\n 6. Whether or not requests for CA certificates are allowed.', args=[d.arg(name='signerName', type=d.T.string)]), + withSignerName(signerName): { spec+: { signerName: signerName } }, + '#withUid':: d.fn(help='uid contains the uid of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUsages':: d.fn(help='usages specifies a set of key usages requested in the issued certificate.\n\nRequests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth".\n\nRequests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth".\n\nValid values are:\n "signing", "digital signature", "content commitment",\n "key encipherment", "key agreement", "data encipherment",\n "cert sign", "crl sign", "encipher only", "decipher only", "any",\n "server auth", "client auth",\n "code signing", "email protection", "s/mime",\n "ipsec end system", "ipsec tunnel", "ipsec user",\n "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc"', args=[d.arg(name='usages', type=d.T.array)]), + withUsages(usages): { spec+: { usages: if std.isArray(v=usages) then usages else [usages] } }, + '#withUsagesMixin':: d.fn(help='usages specifies a set of key usages requested in the issued certificate.\n\nRequests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth".\n\nRequests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth".\n\nValid values are:\n "signing", "digital signature", "content commitment",\n "key encipherment", "key agreement", "data encipherment",\n "cert sign", "crl sign", "encipher only", "decipher only", "any",\n "server auth", "client auth",\n "code signing", "email protection", "s/mime",\n "ipsec end system", "ipsec tunnel", "ipsec user",\n "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='usages', type=d.T.array)]), + withUsagesMixin(usages): { spec+: { usages+: if std.isArray(v=usages) then usages else [usages] } }, + '#withUsername':: d.fn(help='username contains the name of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { spec+: { username: username } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestCondition.libsonnet new file mode 100644 index 00000000000..49944522e3b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestCondition', url='', help='CertificateSigningRequestCondition describes a condition of a CertificateSigningRequest object'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withLastUpdateTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastUpdateTime', type=d.T.string)]), + withLastUpdateTime(lastUpdateTime): { lastUpdateTime: lastUpdateTime }, + '#withMessage':: d.fn(help='message contains a human readable message with details about the request state', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help='reason indicates a brief reason for the request state', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='type of the condition. Known conditions are "Approved", "Denied", and "Failed".\n\nAn "Approved" condition is added via the /approval subresource, indicating the request was approved and should be issued by the signer.\n\nA "Denied" condition is added via the /approval subresource, indicating the request was denied and should not be issued by the signer.\n\nA "Failed" condition is added via the /status subresource, indicating the signer failed to issue the certificate.\n\nApproved and Denied conditions are mutually exclusive. Approved, Denied, and Failed conditions cannot be removed once added.\n\nOnly one condition of a given type is allowed.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestSpec.libsonnet new file mode 100644 index 00000000000..00502a0d1c7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestSpec.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestSpec', url='', help='CertificateSigningRequestSpec contains the certificate request.'), + '#withExtra':: d.fn(help='extra contains extra attributes of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='extra contains extra attributes of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroups':: d.fn(help='groups contains group membership of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { groups: if std.isArray(v=groups) then groups else [groups] }, + '#withGroupsMixin':: d.fn(help='groups contains group membership of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { groups+: if std.isArray(v=groups) then groups else [groups] }, + '#withRequest':: d.fn(help='request contains an x509 certificate signing request encoded in a "CERTIFICATE REQUEST" PEM block. When serialized as JSON or YAML, the data is additionally base64-encoded.', args=[d.arg(name='request', type=d.T.string)]), + withRequest(request): { request: request }, + '#withSignerName':: d.fn(help='signerName indicates the requested signer, and is a qualified name.\n\nList/watch requests for CertificateSigningRequests can filter on this field using a "spec.signerName=NAME" fieldSelector.\n\nWell-known Kubernetes signers are:\n 1. "kubernetes.io/kube-apiserver-client": issues client certificates that can be used to authenticate to kube-apiserver.\n Requests for this signer are never auto-approved by kube-controller-manager, can be issued by the "csrsigning" controller in kube-controller-manager.\n 2. "kubernetes.io/kube-apiserver-client-kubelet": issues client certificates that kubelets use to authenticate to kube-apiserver.\n Requests for this signer can be auto-approved by the "csrapproving" controller in kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager.\n 3. "kubernetes.io/kubelet-serving" issues serving certificates that kubelets use to serve TLS endpoints, which kube-apiserver can connect to securely.\n Requests for this signer are never auto-approved by kube-controller-manager, and can be issued by the "csrsigning" controller in kube-controller-manager.\n\nMore details are available at https://k8s.io/docs/reference/access-authn-authz/certificate-signing-requests/#kubernetes-signers\n\nCustom signerNames can also be specified. The signer defines:\n 1. Trust distribution: how trust (CA bundles) are distributed.\n 2. Permitted subjects: and behavior when a disallowed subject is requested.\n 3. Required, permitted, or forbidden x509 extensions in the request (including whether subjectAltNames are allowed, which types, restrictions on allowed values) and behavior when a disallowed extension is requested.\n 4. Required, permitted, or forbidden key usages / extended key usages.\n 5. Expiration/certificate lifetime: whether it is fixed by the signer, configurable by the admin.\n 6. Whether or not requests for CA certificates are allowed.', args=[d.arg(name='signerName', type=d.T.string)]), + withSignerName(signerName): { signerName: signerName }, + '#withUid':: d.fn(help='uid contains the uid of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUsages':: d.fn(help='usages specifies a set of key usages requested in the issued certificate.\n\nRequests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth".\n\nRequests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth".\n\nValid values are:\n "signing", "digital signature", "content commitment",\n "key encipherment", "key agreement", "data encipherment",\n "cert sign", "crl sign", "encipher only", "decipher only", "any",\n "server auth", "client auth",\n "code signing", "email protection", "s/mime",\n "ipsec end system", "ipsec tunnel", "ipsec user",\n "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc"', args=[d.arg(name='usages', type=d.T.array)]), + withUsages(usages): { usages: if std.isArray(v=usages) then usages else [usages] }, + '#withUsagesMixin':: d.fn(help='usages specifies a set of key usages requested in the issued certificate.\n\nRequests for TLS client certificates typically request: "digital signature", "key encipherment", "client auth".\n\nRequests for TLS serving certificates typically request: "key encipherment", "digital signature", "server auth".\n\nValid values are:\n "signing", "digital signature", "content commitment",\n "key encipherment", "key agreement", "data encipherment",\n "cert sign", "crl sign", "encipher only", "decipher only", "any",\n "server auth", "client auth",\n "code signing", "email protection", "s/mime",\n "ipsec end system", "ipsec tunnel", "ipsec user",\n "timestamping", "ocsp signing", "microsoft sgc", "netscape sgc"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='usages', type=d.T.array)]), + withUsagesMixin(usages): { usages+: if std.isArray(v=usages) then usages else [usages] }, + '#withUsername':: d.fn(help='username contains the name of the user that created the CertificateSigningRequest. Populated by the API server on creation and immutable.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { username: username }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestStatus.libsonnet new file mode 100644 index 00000000000..99d17401046 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/certificateSigningRequestStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestStatus', url='', help='CertificateSigningRequestStatus contains conditions used to indicate approved/denied/failed status of the request, and the issued certificate.'), + '#withCertificate':: d.fn(help='certificate is populated with an issued certificate by the signer after an Approved condition is present. This field is set via the /status subresource. Once populated, this field is immutable.\n\nIf the certificate signing request is denied, a condition of type "Denied" is added and this field remains empty. If the signer cannot issue the certificate, a condition of type "Failed" is added and this field remains empty.\n\nValidation requirements:\n 1. certificate must contain one or more PEM blocks.\n 2. All PEM blocks must have the "CERTIFICATE" label, contain no headers, and the encoded data\n must be a BER-encoded ASN.1 Certificate structure as described in section 4 of RFC5280.\n 3. Non-PEM content may appear before or after the "CERTIFICATE" PEM blocks and is unvalidated,\n to allow for explanatory text as described in section 5.2 of RFC7468.\n\nIf more than one PEM block is present, and the definition of the requested spec.signerName does not indicate otherwise, the first block is the issued certificate, and subsequent blocks should be treated as intermediate certificates and presented in TLS handshakes.\n\nThe certificate is encoded in PEM format.\n\nWhen serialized as JSON or YAML, the data is additionally base64-encoded, so it consists of:\n\n base64(\n -----BEGIN CERTIFICATE-----\n ...\n -----END CERTIFICATE-----\n )', args=[d.arg(name='certificate', type=d.T.string)]), + withCertificate(certificate): { certificate: certificate }, + '#withConditions':: d.fn(help='conditions applied to the request. Known conditions are "Approved", "Denied", and "Failed".', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='conditions applied to the request. Known conditions are "Approved", "Denied", and "Failed".\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/main.libsonnet new file mode 100644 index 00000000000..2f3a81bd083 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + certificateSigningRequest: (import 'certificateSigningRequest.libsonnet'), + certificateSigningRequestCondition: (import 'certificateSigningRequestCondition.libsonnet'), + certificateSigningRequestSpec: (import 'certificateSigningRequestSpec.libsonnet'), + certificateSigningRequestStatus: (import 'certificateSigningRequestStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequest.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequest.libsonnet new file mode 100644 index 00000000000..4663716bc89 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequest.libsonnet @@ -0,0 +1,79 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequest', url='', help='Describes a certificate signing request'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Certificatesigningrequest', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'certificates.k8s.io/v1beta1', + kind: 'CertificateSigningRequest', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='This information is immutable after the request is created. Only the Request and Usages fields can be set on creation, other fields are derived by Kubernetes and cannot be modified by users.'), + spec: { + '#withExtra':: d.fn(help='Extra information about the requesting user. See user.Info interface for details.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { spec+: { extra: extra } }, + '#withExtraMixin':: d.fn(help='Extra information about the requesting user. See user.Info interface for details.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { spec+: { extra+: extra } }, + '#withGroups':: d.fn(help='Group information about the requesting user. See user.Info interface for details.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { spec+: { groups: if std.isArray(v=groups) then groups else [groups] } }, + '#withGroupsMixin':: d.fn(help='Group information about the requesting user. See user.Info interface for details.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { spec+: { groups+: if std.isArray(v=groups) then groups else [groups] } }, + '#withRequest':: d.fn(help='Base64-encoded PKCS#10 CSR data', args=[d.arg(name='request', type=d.T.string)]), + withRequest(request): { spec+: { request: request } }, + '#withSignerName':: d.fn(help="Requested signer for the request. It is a qualified name in the form: `scope-hostname.io/name`. If empty, it will be defaulted:\n 1. If it's a kubelet client certificate, it is assigned\n 'kubernetes.io/kube-apiserver-client-kubelet'.\n 2. If it's a kubelet serving certificate, it is assigned\n 'kubernetes.io/kubelet-serving'.\n 3. Otherwise, it is assigned 'kubernetes.io/legacy-unknown'.\nDistribution of trust for signers happens out of band. You can select on this field using `spec.signerName`.", args=[d.arg(name='signerName', type=d.T.string)]), + withSignerName(signerName): { spec+: { signerName: signerName } }, + '#withUid':: d.fn(help='UID information about the requesting user. See user.Info interface for details.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { uid: uid } }, + '#withUsages':: d.fn(help='allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12\nValid values are:\n "signing",\n "digital signature",\n "content commitment",\n "key encipherment",\n "key agreement",\n "data encipherment",\n "cert sign",\n "crl sign",\n "encipher only",\n "decipher only",\n "any",\n "server auth",\n "client auth",\n "code signing",\n "email protection",\n "s/mime",\n "ipsec end system",\n "ipsec tunnel",\n "ipsec user",\n "timestamping",\n "ocsp signing",\n "microsoft sgc",\n "netscape sgc"', args=[d.arg(name='usages', type=d.T.array)]), + withUsages(usages): { spec+: { usages: if std.isArray(v=usages) then usages else [usages] } }, + '#withUsagesMixin':: d.fn(help='allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12\nValid values are:\n "signing",\n "digital signature",\n "content commitment",\n "key encipherment",\n "key agreement",\n "data encipherment",\n "cert sign",\n "crl sign",\n "encipher only",\n "decipher only",\n "any",\n "server auth",\n "client auth",\n "code signing",\n "email protection",\n "s/mime",\n "ipsec end system",\n "ipsec tunnel",\n "ipsec user",\n "timestamping",\n "ocsp signing",\n "microsoft sgc",\n "netscape sgc"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='usages', type=d.T.array)]), + withUsagesMixin(usages): { spec+: { usages+: if std.isArray(v=usages) then usages else [usages] } }, + '#withUsername':: d.fn(help='Information about the requesting user. See user.Info interface for details.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { spec+: { username: username } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestCondition.libsonnet new file mode 100644 index 00000000000..6bfd9f6ed4f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestCondition', url='', help=''), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withLastUpdateTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastUpdateTime', type=d.T.string)]), + withLastUpdateTime(lastUpdateTime): { lastUpdateTime: lastUpdateTime }, + '#withMessage':: d.fn(help='human readable message with details about the request state', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help='brief reason for the request state', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='type of the condition. Known conditions include "Approved", "Denied", and "Failed".', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestSpec.libsonnet new file mode 100644 index 00000000000..5ef96961e3e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestSpec.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestSpec', url='', help='This information is immutable after the request is created. Only the Request and Usages fields can be set on creation, other fields are derived by Kubernetes and cannot be modified by users.'), + '#withExtra':: d.fn(help='Extra information about the requesting user. See user.Info interface for details.', args=[d.arg(name='extra', type=d.T.object)]), + withExtra(extra): { extra: extra }, + '#withExtraMixin':: d.fn(help='Extra information about the requesting user. See user.Info interface for details.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='extra', type=d.T.object)]), + withExtraMixin(extra): { extra+: extra }, + '#withGroups':: d.fn(help='Group information about the requesting user. See user.Info interface for details.', args=[d.arg(name='groups', type=d.T.array)]), + withGroups(groups): { groups: if std.isArray(v=groups) then groups else [groups] }, + '#withGroupsMixin':: d.fn(help='Group information about the requesting user. See user.Info interface for details.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='groups', type=d.T.array)]), + withGroupsMixin(groups): { groups+: if std.isArray(v=groups) then groups else [groups] }, + '#withRequest':: d.fn(help='Base64-encoded PKCS#10 CSR data', args=[d.arg(name='request', type=d.T.string)]), + withRequest(request): { request: request }, + '#withSignerName':: d.fn(help="Requested signer for the request. It is a qualified name in the form: `scope-hostname.io/name`. If empty, it will be defaulted:\n 1. If it's a kubelet client certificate, it is assigned\n 'kubernetes.io/kube-apiserver-client-kubelet'.\n 2. If it's a kubelet serving certificate, it is assigned\n 'kubernetes.io/kubelet-serving'.\n 3. Otherwise, it is assigned 'kubernetes.io/legacy-unknown'.\nDistribution of trust for signers happens out of band. You can select on this field using `spec.signerName`.", args=[d.arg(name='signerName', type=d.T.string)]), + withSignerName(signerName): { signerName: signerName }, + '#withUid':: d.fn(help='UID information about the requesting user. See user.Info interface for details.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#withUsages':: d.fn(help='allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12\nValid values are:\n "signing",\n "digital signature",\n "content commitment",\n "key encipherment",\n "key agreement",\n "data encipherment",\n "cert sign",\n "crl sign",\n "encipher only",\n "decipher only",\n "any",\n "server auth",\n "client auth",\n "code signing",\n "email protection",\n "s/mime",\n "ipsec end system",\n "ipsec tunnel",\n "ipsec user",\n "timestamping",\n "ocsp signing",\n "microsoft sgc",\n "netscape sgc"', args=[d.arg(name='usages', type=d.T.array)]), + withUsages(usages): { usages: if std.isArray(v=usages) then usages else [usages] }, + '#withUsagesMixin':: d.fn(help='allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3\n https://tools.ietf.org/html/rfc5280#section-4.2.1.12\nValid values are:\n "signing",\n "digital signature",\n "content commitment",\n "key encipherment",\n "key agreement",\n "data encipherment",\n "cert sign",\n "crl sign",\n "encipher only",\n "decipher only",\n "any",\n "server auth",\n "client auth",\n "code signing",\n "email protection",\n "s/mime",\n "ipsec end system",\n "ipsec tunnel",\n "ipsec user",\n "timestamping",\n "ocsp signing",\n "microsoft sgc",\n "netscape sgc"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='usages', type=d.T.array)]), + withUsagesMixin(usages): { usages+: if std.isArray(v=usages) then usages else [usages] }, + '#withUsername':: d.fn(help='Information about the requesting user. See user.Info interface for details.', args=[d.arg(name='username', type=d.T.string)]), + withUsername(username): { username: username }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestStatus.libsonnet new file mode 100644 index 00000000000..26c04d73556 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/certificateSigningRequestStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='certificateSigningRequestStatus', url='', help=''), + '#withCertificate':: d.fn(help='If request was approved, the controller will place the issued certificate here.', args=[d.arg(name='certificate', type=d.T.string)]), + withCertificate(certificate): { certificate: certificate }, + '#withConditions':: d.fn(help='Conditions applied to the request, such as approval or denial.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='Conditions applied to the request, such as approval or denial.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/main.libsonnet new file mode 100644 index 00000000000..0e09e4b298a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/certificates/v1beta1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + certificateSigningRequest: (import 'certificateSigningRequest.libsonnet'), + certificateSigningRequestCondition: (import 'certificateSigningRequestCondition.libsonnet'), + certificateSigningRequestSpec: (import 'certificateSigningRequestSpec.libsonnet'), + certificateSigningRequestStatus: (import 'certificateSigningRequestStatus.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/main.libsonnet new file mode 100644 index 00000000000..9d8b2649c6c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='coordination', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/lease.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/lease.libsonnet new file mode 100644 index 00000000000..2eb3e220594 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/lease.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='lease', url='', help='Lease defines a lease concept.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Lease', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'coordination.k8s.io/v1', + kind: 'Lease', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='LeaseSpec is a specification of a Lease.'), + spec: { + '#withAcquireTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='acquireTime', type=d.T.string)]), + withAcquireTime(acquireTime): { spec+: { acquireTime: acquireTime } }, + '#withHolderIdentity':: d.fn(help='holderIdentity contains the identity of the holder of a current lease.', args=[d.arg(name='holderIdentity', type=d.T.string)]), + withHolderIdentity(holderIdentity): { spec+: { holderIdentity: holderIdentity } }, + '#withLeaseDurationSeconds':: d.fn(help='leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime.', args=[d.arg(name='leaseDurationSeconds', type=d.T.integer)]), + withLeaseDurationSeconds(leaseDurationSeconds): { spec+: { leaseDurationSeconds: leaseDurationSeconds } }, + '#withLeaseTransitions':: d.fn(help='leaseTransitions is the number of transitions of a lease between holders.', args=[d.arg(name='leaseTransitions', type=d.T.integer)]), + withLeaseTransitions(leaseTransitions): { spec+: { leaseTransitions: leaseTransitions } }, + '#withRenewTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='renewTime', type=d.T.string)]), + withRenewTime(renewTime): { spec+: { renewTime: renewTime } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/leaseSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/leaseSpec.libsonnet new file mode 100644 index 00000000000..ee018081058 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/leaseSpec.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='leaseSpec', url='', help='LeaseSpec is a specification of a Lease.'), + '#withAcquireTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='acquireTime', type=d.T.string)]), + withAcquireTime(acquireTime): { acquireTime: acquireTime }, + '#withHolderIdentity':: d.fn(help='holderIdentity contains the identity of the holder of a current lease.', args=[d.arg(name='holderIdentity', type=d.T.string)]), + withHolderIdentity(holderIdentity): { holderIdentity: holderIdentity }, + '#withLeaseDurationSeconds':: d.fn(help='leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime.', args=[d.arg(name='leaseDurationSeconds', type=d.T.integer)]), + withLeaseDurationSeconds(leaseDurationSeconds): { leaseDurationSeconds: leaseDurationSeconds }, + '#withLeaseTransitions':: d.fn(help='leaseTransitions is the number of transitions of a lease between holders.', args=[d.arg(name='leaseTransitions', type=d.T.integer)]), + withLeaseTransitions(leaseTransitions): { leaseTransitions: leaseTransitions }, + '#withRenewTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='renewTime', type=d.T.string)]), + withRenewTime(renewTime): { renewTime: renewTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/main.libsonnet new file mode 100644 index 00000000000..882a9e055ac --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + lease: (import 'lease.libsonnet'), + leaseSpec: (import 'leaseSpec.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/lease.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/lease.libsonnet new file mode 100644 index 00000000000..f3a9d0fdc06 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/lease.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='lease', url='', help='Lease defines a lease concept.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Lease', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'coordination.k8s.io/v1beta1', + kind: 'Lease', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='LeaseSpec is a specification of a Lease.'), + spec: { + '#withAcquireTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='acquireTime', type=d.T.string)]), + withAcquireTime(acquireTime): { spec+: { acquireTime: acquireTime } }, + '#withHolderIdentity':: d.fn(help='holderIdentity contains the identity of the holder of a current lease.', args=[d.arg(name='holderIdentity', type=d.T.string)]), + withHolderIdentity(holderIdentity): { spec+: { holderIdentity: holderIdentity } }, + '#withLeaseDurationSeconds':: d.fn(help='leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime.', args=[d.arg(name='leaseDurationSeconds', type=d.T.integer)]), + withLeaseDurationSeconds(leaseDurationSeconds): { spec+: { leaseDurationSeconds: leaseDurationSeconds } }, + '#withLeaseTransitions':: d.fn(help='leaseTransitions is the number of transitions of a lease between holders.', args=[d.arg(name='leaseTransitions', type=d.T.integer)]), + withLeaseTransitions(leaseTransitions): { spec+: { leaseTransitions: leaseTransitions } }, + '#withRenewTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='renewTime', type=d.T.string)]), + withRenewTime(renewTime): { spec+: { renewTime: renewTime } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/leaseSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/leaseSpec.libsonnet new file mode 100644 index 00000000000..ee018081058 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/leaseSpec.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='leaseSpec', url='', help='LeaseSpec is a specification of a Lease.'), + '#withAcquireTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='acquireTime', type=d.T.string)]), + withAcquireTime(acquireTime): { acquireTime: acquireTime }, + '#withHolderIdentity':: d.fn(help='holderIdentity contains the identity of the holder of a current lease.', args=[d.arg(name='holderIdentity', type=d.T.string)]), + withHolderIdentity(holderIdentity): { holderIdentity: holderIdentity }, + '#withLeaseDurationSeconds':: d.fn(help='leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime.', args=[d.arg(name='leaseDurationSeconds', type=d.T.integer)]), + withLeaseDurationSeconds(leaseDurationSeconds): { leaseDurationSeconds: leaseDurationSeconds }, + '#withLeaseTransitions':: d.fn(help='leaseTransitions is the number of transitions of a lease between holders.', args=[d.arg(name='leaseTransitions', type=d.T.integer)]), + withLeaseTransitions(leaseTransitions): { leaseTransitions: leaseTransitions }, + '#withRenewTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='renewTime', type=d.T.string)]), + withRenewTime(renewTime): { renewTime: renewTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/main.libsonnet new file mode 100644 index 00000000000..12f30209845 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/coordination/v1beta1/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + lease: (import 'lease.libsonnet'), + leaseSpec: (import 'leaseSpec.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/main.libsonnet new file mode 100644 index 00000000000..54730e54cbe --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='core', url='', help=''), + v1: (import 'v1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/affinity.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/affinity.libsonnet new file mode 100644 index 00000000000..5ccef20392c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/affinity.libsonnet @@ -0,0 +1,42 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='affinity', url='', help='Affinity is a group of affinity scheduling rules.'), + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/attachedVolume.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/attachedVolume.libsonnet new file mode 100644 index 00000000000..d7b58684fd9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/attachedVolume.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='attachedVolume', url='', help='AttachedVolume describes a volume attached to a node'), + '#withDevicePath':: d.fn(help='DevicePath represents the device path where the volume should be available', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='Name of the attached volume', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/awsElasticBlockStoreVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/awsElasticBlockStoreVolumeSource.libsonnet new file mode 100644 index 00000000000..fd46e6b9634 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/awsElasticBlockStoreVolumeSource.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='awsElasticBlockStoreVolumeSource', url='', help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { partition: partition }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { volumeID: volumeID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureDiskVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureDiskVolumeSource.libsonnet new file mode 100644 index 00000000000..d64dc07e578 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureDiskVolumeSource.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='azureDiskVolumeSource', url='', help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { cachingMode: cachingMode }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { diskName: diskName }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { diskURI: diskURI }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFilePersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFilePersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..1fa416dd6fb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFilePersistentVolumeSource.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='azureFilePersistentVolumeSource', url='', help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { secretNamespace: secretNamespace }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { shareName: shareName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFileVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFileVolumeSource.libsonnet new file mode 100644 index 00000000000..532767aa31a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/azureFileVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='azureFileVolumeSource', url='', help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { shareName: shareName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/binding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/binding.libsonnet new file mode 100644 index 00000000000..457ff30051f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/binding.libsonnet @@ -0,0 +1,73 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='binding', url='', help='Binding ties one object to another; for example, a pod is bound to a node by a scheduler. Deprecated in 1.7, please use the bindings subresource of pods instead.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Binding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Binding', + } + self.metadata.withName(name=name), + '#target':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + target: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { target+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { target+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { target+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { target+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { target+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { target+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { target+: { uid: uid } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/capabilities.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/capabilities.libsonnet new file mode 100644 index 00000000000..47ab90775ef --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/capabilities.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='capabilities', url='', help='Adds and removes POSIX capabilities from running containers.'), + '#withAdd':: d.fn(help='Added capabilities', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { add: if std.isArray(v=add) then add else [add] }, + '#withAddMixin':: d.fn(help='Added capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { add+: if std.isArray(v=add) then add else [add] }, + '#withDrop':: d.fn(help='Removed capabilities', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { drop: if std.isArray(v=drop) then drop else [drop] }, + '#withDropMixin':: d.fn(help='Removed capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { drop+: if std.isArray(v=drop) then drop else [drop] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..649c1266054 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSPersistentVolumeSource.libsonnet @@ -0,0 +1,25 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cephFSPersistentVolumeSource', url='', help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { monitors: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { monitors+: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { secretFile: secretFile }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSVolumeSource.libsonnet new file mode 100644 index 00000000000..e1c6165fbe8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cephFSVolumeSource.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cephFSVolumeSource', url='', help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { monitors: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { monitors+: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { secretFile: secretFile }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..81fda33c0dc --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderPersistentVolumeSource.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cinderPersistentVolumeSource', url='', help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { volumeID: volumeID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderVolumeSource.libsonnet new file mode 100644 index 00000000000..037f276d41a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/cinderVolumeSource.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='cinderVolumeSource', url='', help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { volumeID: volumeID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/clientIPConfig.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/clientIPConfig.libsonnet new file mode 100644 index 00000000000..8300076641d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/clientIPConfig.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clientIPConfig', url='', help='ClientIPConfig represents the configurations of Client IP based session affinity.'), + '#withTimeoutSeconds':: d.fn(help='timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours).', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentCondition.libsonnet new file mode 100644 index 00000000000..dec59d66ddb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentCondition.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='componentCondition', url='', help='Information about the condition of a component.'), + '#withError':: d.fn(help='Condition error code for a component. For example, a health check error code.', args=[d.arg(name='err', type=d.T.string)]), + withError(err): { 'error': err }, + '#withMessage':: d.fn(help='Message about the condition for a component. For example, information about a health check.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withType':: d.fn(help='Type of condition for a component. Valid value: "Healthy"', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentStatus.libsonnet new file mode 100644 index 00000000000..aab5b89e4d0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/componentStatus.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='componentStatus', url='', help='ComponentStatus (and ComponentStatusList) holds the cluster validation info. Deprecated: This API is deprecated in v1.19+'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Componentstatus', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'ComponentStatus', + } + self.metadata.withName(name=name), + '#withConditions':: d.fn(help='List of component conditions observed', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='List of component conditions observed\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMap.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMap.libsonnet new file mode 100644 index 00000000000..8f4d8f60ad5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMap.libsonnet @@ -0,0 +1,66 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMap', url='', help='ConfigMap holds configuration data for pods to consume.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Configmap', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'ConfigMap', + } + self.metadata.withName(name=name), + '#withBinaryData':: d.fn(help="BinaryData contains the binary data. Each key must consist of alphanumeric characters, '-', '_' or '.'. BinaryData can contain byte sequences that are not in the UTF-8 range. The keys stored in BinaryData must not overlap with the ones in the Data field, this is enforced during validation process. Using this field will require 1.10+ apiserver and kubelet.", args=[d.arg(name='binaryData', type=d.T.object)]), + withBinaryData(binaryData): { binaryData: binaryData }, + '#withBinaryDataMixin':: d.fn(help="BinaryData contains the binary data. Each key must consist of alphanumeric characters, '-', '_' or '.'. BinaryData can contain byte sequences that are not in the UTF-8 range. The keys stored in BinaryData must not overlap with the ones in the Data field, this is enforced during validation process. Using this field will require 1.10+ apiserver and kubelet.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='binaryData', type=d.T.object)]), + withBinaryDataMixin(binaryData): { binaryData+: binaryData }, + '#withData':: d.fn(help="Data contains the configuration data. Each key must consist of alphanumeric characters, '-', '_' or '.'. Values with non-UTF-8 byte sequences must use the BinaryData field. The keys stored in Data must not overlap with the keys in the BinaryData field, this is enforced during validation process.", args=[d.arg(name='data', type=d.T.object)]), + withData(data): { data: data }, + '#withDataMixin':: d.fn(help="Data contains the configuration data. Each key must consist of alphanumeric characters, '-', '_' or '.'. Values with non-UTF-8 byte sequences must use the BinaryData field. The keys stored in Data must not overlap with the keys in the BinaryData field, this is enforced during validation process.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='data', type=d.T.object)]), + withDataMixin(data): { data+: data }, + '#withImmutable':: d.fn(help='Immutable, if set to true, ensures that data stored in the ConfigMap cannot be updated (only object metadata can be modified). If not set to true, the field can be modified at any time. Defaulted to nil. This is a beta field enabled by ImmutableEphemeralVolumes feature gate.', args=[d.arg(name='immutable', type=d.T.boolean)]), + withImmutable(immutable): { immutable: immutable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapEnvSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapEnvSource.libsonnet new file mode 100644 index 00000000000..c5754d68358 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapEnvSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMapEnvSource', url='', help="ConfigMapEnvSource selects a ConfigMap to populate the environment variables with.\n\nThe contents of the target ConfigMap's Data field will represent the key-value pairs as environment variables."), + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapKeySelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapKeySelector.libsonnet new file mode 100644 index 00000000000..cf4ab20a2d5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapKeySelector.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMapKeySelector', url='', help='Selects a key from a ConfigMap.'), + '#withKey':: d.fn(help='The key to select.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapNodeConfigSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapNodeConfigSource.libsonnet new file mode 100644 index 00000000000..49344ee2f63 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapNodeConfigSource.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMapNodeConfigSource', url='', help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { kubeletConfigKey: kubeletConfigKey }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { resourceVersion: resourceVersion }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapProjection.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapProjection.libsonnet new file mode 100644 index 00000000000..fa98ed4eb5f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapProjection.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMapProjection', url='', help="Adapts a ConfigMap into a projected volume.\n\nThe contents of the target ConfigMap's Data field will be presented in a projected volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. Note that this is identical to a configmap volume source without the default mode."), + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapVolumeSource.libsonnet new file mode 100644 index 00000000000..31c8756e48e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/configMapVolumeSource.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='configMapVolumeSource', url='', help="Adapts a ConfigMap into a volume.\n\nThe contents of the target ConfigMap's Data field will be presented in a volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. ConfigMap volumes support ownership management and SELinux relabeling."), + '#withDefaultMode':: d.fn(help='Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { defaultMode: defaultMode }, + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/container.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/container.libsonnet new file mode 100644 index 00000000000..ae94faedf5d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/container.libsonnet @@ -0,0 +1,311 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='container', url='', help='A single application container that you want to run within a pod.'), + '#lifecycle':: d.obj(help='Lifecycle describes actions that the management system should take in response to container lifecycle events. For the PostStart and PreStop lifecycle handlers, management of the container blocks until the action is complete, unless the container process fails, in which case the handler is aborted.'), + lifecycle: { + '#postStart':: d.obj(help='Handler defines a specific action that should be taken'), + postStart: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help='Handler defines a specific action that should be taken'), + preStop: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + livenessProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#readinessProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + readinessProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='SecurityContext holds security configuration that will be applied to a container. Some fields are present in both SecurityContext and PodSecurityContext. When both are set, the values in SecurityContext take precedence.'), + securityContext: { + '#capabilities':: d.obj(help='Adds and removes POSIX capabilities from running containers.'), + capabilities: { + '#withAdd':: d.fn(help='Added capabilities', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='Added capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='Removed capabilities', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='Removed capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false.', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled.', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='Whether this container has a read-only root filesystem. Default is false.', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + startupProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#withArgs':: d.fn(help="Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='List of environment variables to set in the container. Cannot be updated.', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated.', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='List of environment variables to set in the container. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='Docker image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets.', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='List of ports to expose from the container. Exposing a port here gives the system additional information about the network connections a container uses, but is primarily informational. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default "0.0.0.0" address inside a container will be accessible from the network. Cannot be updated.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='List of ports to expose from the container. Exposing a port here gives the system additional information about the network connections a container uses, but is primarily informational. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default "0.0.0.0" address inside a container will be accessible from the network. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withStdin':: d.fn(help='Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false.', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTerminationMessagePath':: d.fn(help="Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated.', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='volumeDevices is the list of block devices to be used by the container.', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='volumeDevices is the list of block devices to be used by the container.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="Pod volumes to mount into the container's filesystem. Cannot be updated.", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="Pod volumes to mount into the container's filesystem. Cannot be updated.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerImage.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerImage.libsonnet new file mode 100644 index 00000000000..6d40fe0ac46 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerImage.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerImage', url='', help='Describe a container image'), + '#withNames':: d.fn(help='Names by which this image is known. e.g. ["k8s.gcr.io/hyperkube:v1.0.7", "dockerhub.io/google_containers/hyperkube:v1.0.7"]', args=[d.arg(name='names', type=d.T.array)]), + withNames(names): { names: if std.isArray(v=names) then names else [names] }, + '#withNamesMixin':: d.fn(help='Names by which this image is known. e.g. ["k8s.gcr.io/hyperkube:v1.0.7", "dockerhub.io/google_containers/hyperkube:v1.0.7"]\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='names', type=d.T.array)]), + withNamesMixin(names): { names+: if std.isArray(v=names) then names else [names] }, + '#withSizeBytes':: d.fn(help='The size of the image in bytes.', args=[d.arg(name='sizeBytes', type=d.T.integer)]), + withSizeBytes(sizeBytes): { sizeBytes: sizeBytes }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerPort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerPort.libsonnet new file mode 100644 index 00000000000..37b08ec37d3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerPort.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerPort', url='', help='ContainerPort represents a network port in a single container.'), + '#withContainerPort':: d.fn(help="Number of port to expose on the pod's IP address. This must be a valid port number, 0 < x < 65536.", args=[d.arg(name='containerPort', type=d.T.integer)]), + withContainerPort(containerPort): { containerPort: containerPort }, + '#withHostIP':: d.fn(help='What host IP to bind the external port to.', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withHostPort':: d.fn(help='Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this.', args=[d.arg(name='hostPort', type=d.T.integer)]), + withHostPort(hostPort): { hostPort: hostPort }, + '#withName':: d.fn(help='If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProtocol':: d.fn(help='Protocol for port. Must be UDP, TCP, or SCTP. Defaults to "TCP".', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerState.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerState.libsonnet new file mode 100644 index 00000000000..d2efec84776 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerState.libsonnet @@ -0,0 +1,35 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerState', url='', help='ContainerState holds a possible state of container. Only one of its members may be specified. If none of them is specified, the default one is ContainerStateWaiting.'), + '#running':: d.obj(help='ContainerStateRunning is a running state of a container.'), + running: { + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { running+: { startedAt: startedAt } }, + }, + '#terminated':: d.obj(help='ContainerStateTerminated is a terminated state of a container.'), + terminated: { + '#withContainerID':: d.fn(help="Container's ID in the format 'docker://'", args=[d.arg(name='containerID', type=d.T.string)]), + withContainerID(containerID): { terminated+: { containerID: containerID } }, + '#withExitCode':: d.fn(help='Exit status from the last termination of the container', args=[d.arg(name='exitCode', type=d.T.integer)]), + withExitCode(exitCode): { terminated+: { exitCode: exitCode } }, + '#withFinishedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='finishedAt', type=d.T.string)]), + withFinishedAt(finishedAt): { terminated+: { finishedAt: finishedAt } }, + '#withMessage':: d.fn(help='Message regarding the last termination of the container', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { terminated+: { message: message } }, + '#withReason':: d.fn(help='(brief) reason from the last termination of the container', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { terminated+: { reason: reason } }, + '#withSignal':: d.fn(help='Signal from the last termination of the container', args=[d.arg(name='signal', type=d.T.integer)]), + withSignal(signal): { terminated+: { signal: signal } }, + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { terminated+: { startedAt: startedAt } }, + }, + '#waiting':: d.obj(help='ContainerStateWaiting is a waiting state of a container.'), + waiting: { + '#withMessage':: d.fn(help='Message regarding why the container is not yet running.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { waiting+: { message: message } }, + '#withReason':: d.fn(help='(brief) reason the container is not yet running.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { waiting+: { reason: reason } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateRunning.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateRunning.libsonnet new file mode 100644 index 00000000000..5eb949918be --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateRunning.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerStateRunning', url='', help='ContainerStateRunning is a running state of a container.'), + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { startedAt: startedAt }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateTerminated.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateTerminated.libsonnet new file mode 100644 index 00000000000..95d1470e779 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateTerminated.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerStateTerminated', url='', help='ContainerStateTerminated is a terminated state of a container.'), + '#withContainerID':: d.fn(help="Container's ID in the format 'docker://'", args=[d.arg(name='containerID', type=d.T.string)]), + withContainerID(containerID): { containerID: containerID }, + '#withExitCode':: d.fn(help='Exit status from the last termination of the container', args=[d.arg(name='exitCode', type=d.T.integer)]), + withExitCode(exitCode): { exitCode: exitCode }, + '#withFinishedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='finishedAt', type=d.T.string)]), + withFinishedAt(finishedAt): { finishedAt: finishedAt }, + '#withMessage':: d.fn(help='Message regarding the last termination of the container', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help='(brief) reason from the last termination of the container', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withSignal':: d.fn(help='Signal from the last termination of the container', args=[d.arg(name='signal', type=d.T.integer)]), + withSignal(signal): { signal: signal }, + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { startedAt: startedAt }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateWaiting.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateWaiting.libsonnet new file mode 100644 index 00000000000..0b2c1d1be15 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStateWaiting.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerStateWaiting', url='', help='ContainerStateWaiting is a waiting state of a container.'), + '#withMessage':: d.fn(help='Message regarding why the container is not yet running.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help='(brief) reason the container is not yet running.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStatus.libsonnet new file mode 100644 index 00000000000..67d387a0af9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/containerStatus.libsonnet @@ -0,0 +1,84 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='containerStatus', url='', help='ContainerStatus contains details for the current status of this container.'), + '#lastState':: d.obj(help='ContainerState holds a possible state of container. Only one of its members may be specified. If none of them is specified, the default one is ContainerStateWaiting.'), + lastState: { + '#running':: d.obj(help='ContainerStateRunning is a running state of a container.'), + running: { + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { lastState+: { running+: { startedAt: startedAt } } }, + }, + '#terminated':: d.obj(help='ContainerStateTerminated is a terminated state of a container.'), + terminated: { + '#withContainerID':: d.fn(help="Container's ID in the format 'docker://'", args=[d.arg(name='containerID', type=d.T.string)]), + withContainerID(containerID): { lastState+: { terminated+: { containerID: containerID } } }, + '#withExitCode':: d.fn(help='Exit status from the last termination of the container', args=[d.arg(name='exitCode', type=d.T.integer)]), + withExitCode(exitCode): { lastState+: { terminated+: { exitCode: exitCode } } }, + '#withFinishedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='finishedAt', type=d.T.string)]), + withFinishedAt(finishedAt): { lastState+: { terminated+: { finishedAt: finishedAt } } }, + '#withMessage':: d.fn(help='Message regarding the last termination of the container', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { lastState+: { terminated+: { message: message } } }, + '#withReason':: d.fn(help='(brief) reason from the last termination of the container', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { lastState+: { terminated+: { reason: reason } } }, + '#withSignal':: d.fn(help='Signal from the last termination of the container', args=[d.arg(name='signal', type=d.T.integer)]), + withSignal(signal): { lastState+: { terminated+: { signal: signal } } }, + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { lastState+: { terminated+: { startedAt: startedAt } } }, + }, + '#waiting':: d.obj(help='ContainerStateWaiting is a waiting state of a container.'), + waiting: { + '#withMessage':: d.fn(help='Message regarding why the container is not yet running.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { lastState+: { waiting+: { message: message } } }, + '#withReason':: d.fn(help='(brief) reason the container is not yet running.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { lastState+: { waiting+: { reason: reason } } }, + }, + }, + '#state':: d.obj(help='ContainerState holds a possible state of container. Only one of its members may be specified. If none of them is specified, the default one is ContainerStateWaiting.'), + state: { + '#running':: d.obj(help='ContainerStateRunning is a running state of a container.'), + running: { + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { state+: { running+: { startedAt: startedAt } } }, + }, + '#terminated':: d.obj(help='ContainerStateTerminated is a terminated state of a container.'), + terminated: { + '#withContainerID':: d.fn(help="Container's ID in the format 'docker://'", args=[d.arg(name='containerID', type=d.T.string)]), + withContainerID(containerID): { state+: { terminated+: { containerID: containerID } } }, + '#withExitCode':: d.fn(help='Exit status from the last termination of the container', args=[d.arg(name='exitCode', type=d.T.integer)]), + withExitCode(exitCode): { state+: { terminated+: { exitCode: exitCode } } }, + '#withFinishedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='finishedAt', type=d.T.string)]), + withFinishedAt(finishedAt): { state+: { terminated+: { finishedAt: finishedAt } } }, + '#withMessage':: d.fn(help='Message regarding the last termination of the container', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { state+: { terminated+: { message: message } } }, + '#withReason':: d.fn(help='(brief) reason from the last termination of the container', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { state+: { terminated+: { reason: reason } } }, + '#withSignal':: d.fn(help='Signal from the last termination of the container', args=[d.arg(name='signal', type=d.T.integer)]), + withSignal(signal): { state+: { terminated+: { signal: signal } } }, + '#withStartedAt':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startedAt', type=d.T.string)]), + withStartedAt(startedAt): { state+: { terminated+: { startedAt: startedAt } } }, + }, + '#waiting':: d.obj(help='ContainerStateWaiting is a waiting state of a container.'), + waiting: { + '#withMessage':: d.fn(help='Message regarding why the container is not yet running.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { state+: { waiting+: { message: message } } }, + '#withReason':: d.fn(help='(brief) reason the container is not yet running.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { state+: { waiting+: { reason: reason } } }, + }, + }, + '#withContainerID':: d.fn(help="Container's ID in the format 'docker://'.", args=[d.arg(name='containerID', type=d.T.string)]), + withContainerID(containerID): { containerID: containerID }, + '#withImage':: d.fn(help='The image the container is running. More info: https://kubernetes.io/docs/concepts/containers/images', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImageID':: d.fn(help="ImageID of the container's image.", args=[d.arg(name='imageID', type=d.T.string)]), + withImageID(imageID): { imageID: imageID }, + '#withName':: d.fn(help='This must be a DNS_LABEL. Each container in a pod must have a unique name. Cannot be updated.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReady':: d.fn(help='Specifies whether the container has passed its readiness probe.', args=[d.arg(name='ready', type=d.T.boolean)]), + withReady(ready): { ready: ready }, + '#withRestartCount':: d.fn(help='The number of times the container has been restarted, currently based on the number of dead containers that have not yet been removed. Note that this is calculated from dead containers. But those containers are subject to garbage collection. This value will get capped at 5 by GC.', args=[d.arg(name='restartCount', type=d.T.integer)]), + withRestartCount(restartCount): { restartCount: restartCount }, + '#withStarted':: d.fn(help='Specifies whether the container has passed its startup probe. Initialized as false, becomes true after startupProbe is considered successful. Resets to false when the container is restarted, or if kubelet loses state temporarily. Is always true when no startupProbe is defined.', args=[d.arg(name='started', type=d.T.boolean)]), + withStarted(started): { started: started }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..af740572f9c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiPersistentVolumeSource.libsonnet @@ -0,0 +1,46 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiPersistentVolumeSource', url='', help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { controllerExpandSecretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { controllerExpandSecretRef+: { namespace: namespace } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { controllerPublishSecretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { controllerPublishSecretRef+: { namespace: namespace } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { nodePublishSecretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { nodePublishSecretRef+: { namespace: namespace } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { nodeStageSecretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { nodeStageSecretRef+: { namespace: namespace } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { driver: driver }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { volumeAttributes: volumeAttributes }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { volumeAttributes+: volumeAttributes }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { volumeHandle: volumeHandle }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiVolumeSource.libsonnet new file mode 100644 index 00000000000..6291fa787e6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/csiVolumeSource.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiVolumeSource', url='', help='Represents a source location of a volume to mount, managed by an external CSI driver'), + '#nodePublishSecretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { nodePublishSecretRef+: { name: name } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { driver: driver }, + '#withFsType':: d.fn(help='Filesystem type to mount. Ex. "ext4", "xfs", "ntfs". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Specifies a read-only configuration for the volume. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeAttributes':: d.fn(help="VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { volumeAttributes: volumeAttributes }, + '#withVolumeAttributesMixin':: d.fn(help="VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { volumeAttributes+: volumeAttributes }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/daemonEndpoint.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/daemonEndpoint.libsonnet new file mode 100644 index 00000000000..229f9950cf5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/daemonEndpoint.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='daemonEndpoint', url='', help='DaemonEndpoint contains information about a single Daemon endpoint.'), + '#withPort':: d.fn(help='Port number of the given endpoint.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIProjection.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIProjection.libsonnet new file mode 100644 index 00000000000..2981dd5858b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIProjection.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='downwardAPIProjection', url='', help='Represents downward API info for projecting into a projected volume. Note that this is identical to a downwardAPI volume source without the default mode.'), + '#withItems':: d.fn(help='Items is a list of DownwardAPIVolume file', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help='Items is a list of DownwardAPIVolume file\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeFile.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeFile.libsonnet new file mode 100644 index 00000000000..13e0b54f025 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeFile.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='downwardAPIVolumeFile', url='', help='DownwardAPIVolumeFile represents information to create the file containing the pod field'), + '#fieldRef':: d.obj(help='ObjectFieldSelector selects an APIVersioned field of an object.'), + fieldRef: { + '#withApiVersion':: d.fn(help='Version of the schema the FieldPath is written in terms of, defaults to "v1".', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='Path of the field to select in the specified API version.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='ResourceFieldSelector represents container resources (cpu, memory) and their output format'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='Container name: required for volumes, optional for env vars', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='divisor', type=d.T.string)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='Required: resource to select', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#withMode':: d.fn(help='Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeSource.libsonnet new file mode 100644 index 00000000000..5791b13b019 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/downwardAPIVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='downwardAPIVolumeSource', url='', help='DownwardAPIVolumeSource represents a volume containing downward API info. Downward API volumes support ownership management and SELinux relabeling.'), + '#withDefaultMode':: d.fn(help='Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { defaultMode: defaultMode }, + '#withItems':: d.fn(help='Items is a list of downward API volume file', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help='Items is a list of downward API volume file\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/emptyDirVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/emptyDirVolumeSource.libsonnet new file mode 100644 index 00000000000..2c99ed71080 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/emptyDirVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='emptyDirVolumeSource', url='', help='Represents an empty directory for a pod. Empty directory volumes support ownership management and SELinux relabeling.'), + '#withMedium':: d.fn(help="What type of storage medium should back this directory. The default is '' which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { medium: medium }, + '#withSizeLimit':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='sizeLimit', type=d.T.string)]), + withSizeLimit(sizeLimit): { sizeLimit: sizeLimit }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointAddress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointAddress.libsonnet new file mode 100644 index 00000000000..1d47bffba6b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointAddress.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointAddress', url='', help='EndpointAddress is a tuple that describes single IP address.'), + '#targetRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + targetRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { targetRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { targetRef+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { targetRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { targetRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { targetRef+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { targetRef+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { targetRef+: { uid: uid } }, + }, + '#withHostname':: d.fn(help='The Hostname of this endpoint', args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { hostname: hostname }, + '#withIp':: d.fn(help='The IP of this endpoint. May not be loopback (127.0.0.0/8), link-local (169.254.0.0/16), or link-local multicast ((224.0.0.0/24). IPv6 is also accepted but not fully supported on all platforms. Also, certain kubernetes components, like kube-proxy, are not IPv6 ready.', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + '#withNodeName':: d.fn(help='Optional: Node hosting this endpoint. This can be used to determine endpoints local to a node.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { nodeName: nodeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointPort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointPort.libsonnet new file mode 100644 index 00000000000..0571ff1f314 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointPort.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointPort', url='', help='EndpointPort is a tuple that describes a single port.'), + '#withAppProtocol':: d.fn(help='The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol. This is a beta field that is guarded by the ServiceAppProtocol feature gate and enabled by default.', args=[d.arg(name='appProtocol', type=d.T.string)]), + withAppProtocol(appProtocol): { appProtocol: appProtocol }, + '#withName':: d.fn(help="The name of this port. This must match the 'name' field in the corresponding ServicePort. Must be a DNS_LABEL. Optional only if one port is defined.", args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPort':: d.fn(help='The port number of the endpoint.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#withProtocol':: d.fn(help='The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP.', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointSubset.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointSubset.libsonnet new file mode 100644 index 00000000000..36d2c1db8a5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpointSubset.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointSubset', url='', help='EndpointSubset is a group of addresses with a common set of ports. The expanded set of endpoints is the Cartesian product of Addresses x Ports. For example, given:\n {\n Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}],\n Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}]\n }\nThe resulting set of endpoints can be viewed as:\n a: [ 10.10.1.1:8675, 10.10.2.2:8675 ],\n b: [ 10.10.1.1:309, 10.10.2.2:309 ]'), + '#withAddresses':: d.fn(help='IP addresses which offer the related ports that are marked as ready. These endpoints should be considered safe for load balancers and clients to utilize.', args=[d.arg(name='addresses', type=d.T.array)]), + withAddresses(addresses): { addresses: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withAddressesMixin':: d.fn(help='IP addresses which offer the related ports that are marked as ready. These endpoints should be considered safe for load balancers and clients to utilize.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='addresses', type=d.T.array)]), + withAddressesMixin(addresses): { addresses+: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withNotReadyAddresses':: d.fn(help='IP addresses which offer the related ports but are not currently marked as ready because they have not yet finished starting, have recently failed a readiness check, or have recently failed a liveness check.', args=[d.arg(name='notReadyAddresses', type=d.T.array)]), + withNotReadyAddresses(notReadyAddresses): { notReadyAddresses: if std.isArray(v=notReadyAddresses) then notReadyAddresses else [notReadyAddresses] }, + '#withNotReadyAddressesMixin':: d.fn(help='IP addresses which offer the related ports but are not currently marked as ready because they have not yet finished starting, have recently failed a readiness check, or have recently failed a liveness check.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='notReadyAddresses', type=d.T.array)]), + withNotReadyAddressesMixin(notReadyAddresses): { notReadyAddresses+: if std.isArray(v=notReadyAddresses) then notReadyAddresses else [notReadyAddresses] }, + '#withPorts':: d.fn(help='Port numbers available on the related IP addresses.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='Port numbers available on the related IP addresses.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpoints.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpoints.libsonnet new file mode 100644 index 00000000000..d596e259c78 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/endpoints.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpoints', url='', help='Endpoints is a collection of endpoints that implement the actual service. Example:\n Name: "mysvc",\n Subsets: [\n {\n Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}],\n Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}]\n },\n {\n Addresses: [{"ip": "10.10.3.3"}],\n Ports: [{"name": "a", "port": 93}, {"name": "b", "port": 76}]\n },\n ]'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Endpoints', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Endpoints', + } + self.metadata.withName(name=name), + '#withSubsets':: d.fn(help='The set of all endpoints is the union of all subsets. Addresses are placed into subsets according to the IPs they share. A single address with multiple ports, some of which are ready and some of which are not (because they come from different containers) will result in the address being displayed in different subsets for the different ports. No address will appear in both Addresses and NotReadyAddresses in the same subset. Sets of addresses and ports that comprise a service.', args=[d.arg(name='subsets', type=d.T.array)]), + withSubsets(subsets): { subsets: if std.isArray(v=subsets) then subsets else [subsets] }, + '#withSubsetsMixin':: d.fn(help='The set of all endpoints is the union of all subsets. Addresses are placed into subsets according to the IPs they share. A single address with multiple ports, some of which are ready and some of which are not (because they come from different containers) will result in the address being displayed in different subsets for the different ports. No address will appear in both Addresses and NotReadyAddresses in the same subset. Sets of addresses and ports that comprise a service.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subsets', type=d.T.array)]), + withSubsetsMixin(subsets): { subsets+: if std.isArray(v=subsets) then subsets else [subsets] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envFromSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envFromSource.libsonnet new file mode 100644 index 00000000000..646ed6cd270 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envFromSource.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='envFromSource', url='', help='EnvFromSource represents the source of a set of ConfigMaps'), + '#configMapRef':: d.obj(help="ConfigMapEnvSource selects a ConfigMap to populate the environment variables with.\n\nThe contents of the target ConfigMap's Data field will represent the key-value pairs as environment variables."), + configMapRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapRef+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapRef+: { optional: optional } }, + }, + '#secretRef':: d.obj(help="SecretEnvSource selects a Secret to populate the environment variables with.\n\nThe contents of the target Secret's Data field will represent the key-value pairs as environment variables."), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the Secret must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretRef+: { optional: optional } }, + }, + '#withPrefix':: d.fn(help='An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER.', args=[d.arg(name='prefix', type=d.T.string)]), + withPrefix(prefix): { prefix: prefix }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVar.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVar.libsonnet new file mode 100644 index 00000000000..510165960df --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVar.libsonnet @@ -0,0 +1,47 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='envVar', url='', help='EnvVar represents an environment variable present in a Container.'), + '#valueFrom':: d.obj(help='EnvVarSource represents a source for the value of an EnvVar.'), + valueFrom: { + '#configMapKeyRef':: d.obj(help='Selects a key from a ConfigMap.'), + configMapKeyRef: { + '#withKey':: d.fn(help='The key to select.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } }, + }, + '#fieldRef':: d.obj(help='ObjectFieldSelector selects an APIVersioned field of an object.'), + fieldRef: { + '#withApiVersion':: d.fn(help='Version of the schema the FieldPath is written in terms of, defaults to "v1".', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='Path of the field to select in the specified API version.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } }, + }, + '#resourceFieldRef':: d.obj(help='ResourceFieldSelector represents container resources (cpu, memory) and their output format'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='Container name: required for volumes, optional for env vars', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } }, + '#withDivisor':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='divisor', type=d.T.string)]), + withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } }, + '#withResource':: d.fn(help='Required: resource to select', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } }, + }, + '#secretKeyRef':: d.obj(help='SecretKeySelector selects a key of a Secret.'), + secretKeyRef: { + '#withKey':: d.fn(help='The key of the secret to select from. Must be a valid secret key.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { secretKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='Specify whether the Secret or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } }, + }, + }, + '#withName':: d.fn(help='Name of the environment variable. Must be a C_IDENTIFIER.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='Variable references $(VAR_NAME) are expanded using the previous defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to "".', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVarSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVarSource.libsonnet new file mode 100644 index 00000000000..e0fd45ac570 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/envVarSource.libsonnet @@ -0,0 +1,40 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='envVarSource', url='', help='EnvVarSource represents a source for the value of an EnvVar.'), + '#configMapKeyRef':: d.obj(help='Selects a key from a ConfigMap.'), + configMapKeyRef: { + '#withKey':: d.fn(help='The key to select.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { configMapKeyRef+: { key: key } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapKeyRef+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapKeyRef+: { optional: optional } }, + }, + '#fieldRef':: d.obj(help='ObjectFieldSelector selects an APIVersioned field of an object.'), + fieldRef: { + '#withApiVersion':: d.fn(help='Version of the schema the FieldPath is written in terms of, defaults to "v1".', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='Path of the field to select in the specified API version.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='ResourceFieldSelector represents container resources (cpu, memory) and their output format'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='Container name: required for volumes, optional for env vars', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='divisor', type=d.T.string)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='Required: resource to select', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#secretKeyRef':: d.obj(help='SecretKeySelector selects a key of a Secret.'), + secretKeyRef: { + '#withKey':: d.fn(help='The key of the secret to select from. Must be a valid secret key.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { secretKeyRef+: { key: key } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretKeyRef+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the Secret or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretKeyRef+: { optional: optional } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralContainer.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralContainer.libsonnet new file mode 100644 index 00000000000..6540a5e1d54 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralContainer.libsonnet @@ -0,0 +1,313 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ephemeralContainer', url='', help="An EphemeralContainer is a container that may be added temporarily to an existing pod for user-initiated activities such as debugging. Ephemeral containers have no resource or scheduling guarantees, and they will not be restarted when they exit or when a pod is removed or restarted. If an ephemeral container causes a pod to exceed its resource allocation, the pod may be evicted. Ephemeral containers may not be added by directly updating the pod spec. They must be added via the pod's ephemeralcontainers subresource, and they will appear in the pod spec once added. This is an alpha feature enabled by the EphemeralContainers feature flag."), + '#lifecycle':: d.obj(help='Lifecycle describes actions that the management system should take in response to container lifecycle events. For the PostStart and PreStop lifecycle handlers, management of the container blocks until the action is complete, unless the container process fails, in which case the handler is aborted.'), + lifecycle: { + '#postStart':: d.obj(help='Handler defines a specific action that should be taken'), + postStart: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help='Handler defines a specific action that should be taken'), + preStop: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + livenessProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#readinessProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + readinessProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='SecurityContext holds security configuration that will be applied to a container. Some fields are present in both SecurityContext and PodSecurityContext. When both are set, the values in SecurityContext take precedence.'), + securityContext: { + '#capabilities':: d.obj(help='Adds and removes POSIX capabilities from running containers.'), + capabilities: { + '#withAdd':: d.fn(help='Added capabilities', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='Added capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='Removed capabilities', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='Removed capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false.', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled.', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='Whether this container has a read-only root filesystem. Default is false.', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + startupProbe: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#withArgs':: d.fn(help="Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='List of environment variables to set in the container. Cannot be updated.', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated.', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='List of environment variables to set in the container. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='Docker image name. More info: https://kubernetes.io/docs/concepts/containers/images', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='Name of the ephemeral container specified as a DNS_LABEL. This name must be unique among all containers, init containers and ephemeral containers.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='Ports are not allowed for ephemeral containers.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='Ports are not allowed for ephemeral containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withStdin':: d.fn(help='Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false.', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTargetContainerName':: d.fn(help='If set, the name of the container from PodSpec that this ephemeral container targets. The ephemeral container will be run in the namespaces (IPC, PID, etc) of this container. If not set then the ephemeral container is run in whatever namespaces are shared for the pod. Note that the container runtime must support this feature.', args=[d.arg(name='targetContainerName', type=d.T.string)]), + withTargetContainerName(targetContainerName): { targetContainerName: targetContainerName }, + '#withTerminationMessagePath':: d.fn(help="Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated.', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='volumeDevices is the list of block devices to be used by the container.', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='volumeDevices is the list of block devices to be used by the container.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="Pod volumes to mount into the container's filesystem. Cannot be updated.", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="Pod volumes to mount into the container's filesystem. Cannot be updated.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralVolumeSource.libsonnet new file mode 100644 index 00000000000..3c3df0242f8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/ephemeralVolumeSource.libsonnet @@ -0,0 +1,100 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ephemeralVolumeSource', url='', help='Represents an ephemeral volume that is handled by a normal storage driver.'), + '#volumeClaimTemplate':: d.obj(help='PersistentVolumeClaimTemplate is used to produce PersistentVolumeClaim objects as part of an EphemeralVolumeSource.'), + volumeClaimTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { volumeClaimTemplate+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { volumeClaimTemplate+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { volumeClaimTemplate+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { volumeClaimTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { volumeClaimTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { volumeClaimTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { volumeClaimTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { volumeClaimTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { volumeClaimTemplate+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { volumeClaimTemplate+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { volumeClaimTemplate+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { volumeClaimTemplate+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { volumeClaimTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { volumeClaimTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { volumeClaimTemplate+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { volumeClaimTemplate+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { volumeClaimTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { volumeClaimTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { volumeClaimTemplate+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { volumeClaimTemplate+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { volumeClaimTemplate+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes'), + spec: { + '#dataSource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + dataSource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } }, + }, + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withStorageClassName':: d.fn(help='Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } }, + '#withVolumeName':: d.fn(help='VolumeName is the binding reference to the PersistentVolume backing this claim.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } }, + }, + }, + '#withReadOnly':: d.fn(help='Specifies a read-only configuration for the volume. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/event.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/event.libsonnet new file mode 100644 index 00000000000..539820050ee --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/event.libsonnet @@ -0,0 +1,124 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='event', url='', help='Event is a report of an event somewhere in the cluster.'), + '#involvedObject':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + involvedObject: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { involvedObject+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { involvedObject+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { involvedObject+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { involvedObject+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { involvedObject+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { involvedObject+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { involvedObject+: { uid: uid } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Event', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Event', + } + self.metadata.withName(name=name), + '#related':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + related: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { related+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { related+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { related+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { related+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { related+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { related+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { related+: { uid: uid } }, + }, + '#series':: d.obj(help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.'), + series: { + '#withCount':: d.fn(help='Number of occurrences in this series up to the last heartbeat time', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { series+: { count: count } }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { series+: { lastObservedTime: lastObservedTime } }, + }, + '#source':: d.obj(help='EventSource contains information for an event.'), + source: { + '#withComponent':: d.fn(help='Component from which the event is generated.', args=[d.arg(name='component', type=d.T.string)]), + withComponent(component): { source+: { component: component } }, + '#withHost':: d.fn(help='Node name on which the event is generated.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { source+: { host: host } }, + }, + '#withAction':: d.fn(help='What action was taken/failed regarding to the Regarding object.', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withCount':: d.fn(help='The number of times this event has occurred.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#withEventTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='eventTime', type=d.T.string)]), + withEventTime(eventTime): { eventTime: eventTime }, + '#withFirstTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='firstTimestamp', type=d.T.string)]), + withFirstTimestamp(firstTimestamp): { firstTimestamp: firstTimestamp }, + '#withLastTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTimestamp', type=d.T.string)]), + withLastTimestamp(lastTimestamp): { lastTimestamp: lastTimestamp }, + '#withMessage':: d.fn(help='A human-readable description of the status of this operation.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="This should be a short, machine understandable string that gives the reason for the transition into the object's current status.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withReportingComponent':: d.fn(help='Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`.', args=[d.arg(name='reportingComponent', type=d.T.string)]), + withReportingComponent(reportingComponent): { reportingComponent: reportingComponent }, + '#withReportingInstance':: d.fn(help='ID of the controller instance, e.g. `kubelet-xyzf`.', args=[d.arg(name='reportingInstance', type=d.T.string)]), + withReportingInstance(reportingInstance): { reportingInstance: reportingInstance }, + '#withType':: d.fn(help='Type of this event (Normal, Warning), new types could be added in the future', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSeries.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSeries.libsonnet new file mode 100644 index 00000000000..0832d8f3269 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSeries.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='eventSeries', url='', help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.'), + '#withCount':: d.fn(help='Number of occurrences in this series up to the last heartbeat time', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { lastObservedTime: lastObservedTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSource.libsonnet new file mode 100644 index 00000000000..481fe8e348c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/eventSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='eventSource', url='', help='EventSource contains information for an event.'), + '#withComponent':: d.fn(help='Component from which the event is generated.', args=[d.arg(name='component', type=d.T.string)]), + withComponent(component): { component: component }, + '#withHost':: d.fn(help='Node name on which the event is generated.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/execAction.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/execAction.libsonnet new file mode 100644 index 00000000000..2f7256a74fd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/execAction.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='execAction', url='', help='ExecAction describes a "run in container" action.'), + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/fcVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/fcVolumeSource.libsonnet new file mode 100644 index 00000000000..15792b2ab9c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/fcVolumeSource.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='fcVolumeSource', url='', help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { lun: lun }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { wwids: if std.isArray(v=wwids) then wwids else [wwids] }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { wwids+: if std.isArray(v=wwids) then wwids else [wwids] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..107b8af400b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexPersistentVolumeSource.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flexPersistentVolumeSource', url='', help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { driver: driver }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { options: options }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { options+: options }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexVolumeSource.libsonnet new file mode 100644 index 00000000000..cd42df6d5b1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flexVolumeSource.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flexVolumeSource', url='', help='FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { driver: driver }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { options: options }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { options+: options }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flockerVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flockerVolumeSource.libsonnet new file mode 100644 index 00000000000..c8cd49ded34 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/flockerVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flockerVolumeSource', url='', help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { datasetName: datasetName }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { datasetUUID: datasetUUID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gcePersistentDiskVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gcePersistentDiskVolumeSource.libsonnet new file mode 100644 index 00000000000..c8a80d2a135 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gcePersistentDiskVolumeSource.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='gcePersistentDiskVolumeSource', url='', help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { partition: partition }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { pdName: pdName }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gitRepoVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gitRepoVolumeSource.libsonnet new file mode 100644 index 00000000000..ed2d9709855 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/gitRepoVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='gitRepoVolumeSource', url='', help="Represents a volume that is populated with the contents of a git repository. Git repo volumes do not support ownership management. Git repo volumes support SELinux relabeling.\n\nDEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container."), + '#withDirectory':: d.fn(help="Target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name.", args=[d.arg(name='directory', type=d.T.string)]), + withDirectory(directory): { directory: directory }, + '#withRepository':: d.fn(help='Repository URL', args=[d.arg(name='repository', type=d.T.string)]), + withRepository(repository): { repository: repository }, + '#withRevision':: d.fn(help='Commit hash for the specified revision.', args=[d.arg(name='revision', type=d.T.string)]), + withRevision(revision): { revision: revision }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..e3f38d605ff --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsPersistentVolumeSource.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='glusterfsPersistentVolumeSource', url='', help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { endpoints: endpoints }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { endpointsNamespace: endpointsNamespace }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsVolumeSource.libsonnet new file mode 100644 index 00000000000..bd6e09d877d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/glusterfsVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='glusterfsVolumeSource', url='', help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { endpoints: endpoints }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/handler.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/handler.libsonnet new file mode 100644 index 00000000000..076cbf63651 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/handler.libsonnet @@ -0,0 +1,35 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='handler', url='', help='Handler defines a specific action that should be taken'), + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { exec+: { command: if std.isArray(v=command) then command else [command] } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { exec+: { command+: if std.isArray(v=command) then command else [command] } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { httpGet+: { host: host } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { httpGet+: { path: path } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { httpGet+: { port: port } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { httpGet+: { scheme: scheme } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { tcpSocket+: { host: host } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { tcpSocket+: { port: port } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostAlias.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostAlias.libsonnet new file mode 100644 index 00000000000..bc4a29b331e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostAlias.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='hostAlias', url='', help="HostAlias holds the mapping between IP and hostnames that will be injected as an entry in the pod's hosts file."), + '#withHostnames':: d.fn(help='Hostnames for the above IP address.', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnames(hostnames): { hostnames: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withHostnamesMixin':: d.fn(help='Hostnames for the above IP address.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnamesMixin(hostnames): { hostnames+: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withIp':: d.fn(help='IP address of the host file entry.', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostPathVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostPathVolumeSource.libsonnet new file mode 100644 index 00000000000..488c793ced5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/hostPathVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='hostPathVolumeSource', url='', help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpGetAction.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpGetAction.libsonnet new file mode 100644 index 00000000000..506e743ca85 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpGetAction.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpGetAction', url='', help='HTTPGetAction describes an action based on HTTP Get requests.'), + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { port: port }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { scheme: scheme }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpHeader.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpHeader.libsonnet new file mode 100644 index 00000000000..72edfc05227 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/httpHeader.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpHeader', url='', help='HTTPHeader describes a custom header to be used in HTTP probes'), + '#withName':: d.fn(help='The header field name', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='The header field value', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..8b461b5661a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiPersistentVolumeSource.libsonnet @@ -0,0 +1,35 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='iscsiPersistentVolumeSource', url='', help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { chapAuthDiscovery: chapAuthDiscovery }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { chapAuthSession: chapAuthSession }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { initiatorName: initiatorName }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iqn: iqn }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsiInterface: iscsiInterface }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { lun: lun }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { portals: if std.isArray(v=portals) then portals else [portals] }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { portals+: if std.isArray(v=portals) then portals else [portals] }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { targetPortal: targetPortal }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiVolumeSource.libsonnet new file mode 100644 index 00000000000..d67c131e87d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/iscsiVolumeSource.libsonnet @@ -0,0 +1,33 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='iscsiVolumeSource', url='', help='Represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { chapAuthDiscovery: chapAuthDiscovery }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { chapAuthSession: chapAuthSession }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { initiatorName: initiatorName }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iqn: iqn }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsiInterface: iscsiInterface }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { lun: lun }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { portals: if std.isArray(v=portals) then portals else [portals] }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { portals+: if std.isArray(v=portals) then portals else [portals] }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { targetPortal: targetPortal }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/keyToPath.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/keyToPath.libsonnet new file mode 100644 index 00000000000..5b6e6c7a458 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/keyToPath.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='keyToPath', url='', help='Maps a string key to a path within a volume.'), + '#withKey':: d.fn(help='The key to project.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="The relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/lifecycle.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/lifecycle.libsonnet new file mode 100644 index 00000000000..7fd4b6074c5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/lifecycle.libsonnet @@ -0,0 +1,70 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='lifecycle', url='', help='Lifecycle describes actions that the management system should take in response to container lifecycle events. For the PostStart and PreStop lifecycle handlers, management of the container blocks until the action is complete, unless the container process fails, in which case the handler is aborted.'), + '#postStart':: d.obj(help='Handler defines a specific action that should be taken'), + postStart: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { postStart+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { postStart+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { postStart+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { postStart+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { postStart+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { postStart+: { tcpSocket+: { port: port } } }, + }, + }, + '#preStop':: d.obj(help='Handler defines a specific action that should be taken'), + preStop: { + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { preStop+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { preStop+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { preStop+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { preStop+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { preStop+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { preStop+: { tcpSocket+: { port: port } } }, + }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRange.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRange.libsonnet new file mode 100644 index 00000000000..391482c7b31 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRange.libsonnet @@ -0,0 +1,63 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='limitRange', url='', help='LimitRange sets resource usage limits for each kind of resource in a Namespace.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Limitrange', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'LimitRange', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='LimitRangeSpec defines a min/max usage limit for resources that match on kind.'), + spec: { + '#withLimits':: d.fn(help='Limits is the list of LimitRangeItem objects that are enforced.', args=[d.arg(name='limits', type=d.T.array)]), + withLimits(limits): { spec+: { limits: if std.isArray(v=limits) then limits else [limits] } }, + '#withLimitsMixin':: d.fn(help='Limits is the list of LimitRangeItem objects that are enforced.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.array)]), + withLimitsMixin(limits): { spec+: { limits+: if std.isArray(v=limits) then limits else [limits] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeItem.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeItem.libsonnet new file mode 100644 index 00000000000..27ad7bb2f54 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeItem.libsonnet @@ -0,0 +1,28 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='limitRangeItem', url='', help='LimitRangeItem defines a min/max usage limit for any resource that matches on kind.'), + '#withDefault':: d.fn(help='Default resource requirement limit value by resource name if resource limit is omitted.', args=[d.arg(name='default', type=d.T.object)]), + withDefault(default): { default: default }, + '#withDefaultMixin':: d.fn(help='Default resource requirement limit value by resource name if resource limit is omitted.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='default', type=d.T.object)]), + withDefaultMixin(default): { default+: default }, + '#withDefaultRequest':: d.fn(help='DefaultRequest is the default resource requirement request value by resource name if resource request is omitted.', args=[d.arg(name='defaultRequest', type=d.T.object)]), + withDefaultRequest(defaultRequest): { defaultRequest: defaultRequest }, + '#withDefaultRequestMixin':: d.fn(help='DefaultRequest is the default resource requirement request value by resource name if resource request is omitted.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='defaultRequest', type=d.T.object)]), + withDefaultRequestMixin(defaultRequest): { defaultRequest+: defaultRequest }, + '#withMax':: d.fn(help='Max usage constraints on this kind by resource name.', args=[d.arg(name='max', type=d.T.object)]), + withMax(max): { max: max }, + '#withMaxLimitRequestRatio':: d.fn(help='MaxLimitRequestRatio if specified, the named resource must have a request and limit that are both non-zero where limit divided by request is less than or equal to the enumerated value; this represents the max burst for the named resource.', args=[d.arg(name='maxLimitRequestRatio', type=d.T.object)]), + withMaxLimitRequestRatio(maxLimitRequestRatio): { maxLimitRequestRatio: maxLimitRequestRatio }, + '#withMaxLimitRequestRatioMixin':: d.fn(help='MaxLimitRequestRatio if specified, the named resource must have a request and limit that are both non-zero where limit divided by request is less than or equal to the enumerated value; this represents the max burst for the named resource.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='maxLimitRequestRatio', type=d.T.object)]), + withMaxLimitRequestRatioMixin(maxLimitRequestRatio): { maxLimitRequestRatio+: maxLimitRequestRatio }, + '#withMaxMixin':: d.fn(help='Max usage constraints on this kind by resource name.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='max', type=d.T.object)]), + withMaxMixin(max): { max+: max }, + '#withMin':: d.fn(help='Min usage constraints on this kind by resource name.', args=[d.arg(name='min', type=d.T.object)]), + withMin(min): { min: min }, + '#withMinMixin':: d.fn(help='Min usage constraints on this kind by resource name.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='min', type=d.T.object)]), + withMinMixin(min): { min+: min }, + '#withType':: d.fn(help='Type of resource that this limit applies to.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeSpec.libsonnet new file mode 100644 index 00000000000..bf8d5fe6868 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/limitRangeSpec.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='limitRangeSpec', url='', help='LimitRangeSpec defines a min/max usage limit for resources that match on kind.'), + '#withLimits':: d.fn(help='Limits is the list of LimitRangeItem objects that are enforced.', args=[d.arg(name='limits', type=d.T.array)]), + withLimits(limits): { limits: if std.isArray(v=limits) then limits else [limits] }, + '#withLimitsMixin':: d.fn(help='Limits is the list of LimitRangeItem objects that are enforced.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.array)]), + withLimitsMixin(limits): { limits+: if std.isArray(v=limits) then limits else [limits] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerIngress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerIngress.libsonnet new file mode 100644 index 00000000000..ccf43aa21fa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerIngress.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='loadBalancerIngress', url='', help='LoadBalancerIngress represents the status of a load-balancer ingress point: traffic intended for the service should be sent to an ingress point.'), + '#withHostname':: d.fn(help='Hostname is set for load-balancer ingress points that are DNS based (typically AWS load-balancers)', args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { hostname: hostname }, + '#withIp':: d.fn(help='IP is set for load-balancer ingress points that are IP based (typically GCE or OpenStack load-balancers)', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerStatus.libsonnet new file mode 100644 index 00000000000..b70844b704d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/loadBalancerStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='loadBalancerStatus', url='', help='LoadBalancerStatus represents the status of a load-balancer.'), + '#withIngress':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.', args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { ingress: if std.isArray(v=ingress) then ingress else [ingress] }, + '#withIngressMixin':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { ingress+: if std.isArray(v=ingress) then ingress else [ingress] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localObjectReference.libsonnet new file mode 100644 index 00000000000..894f562249f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localObjectReference.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='localObjectReference', url='', help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localVolumeSource.libsonnet new file mode 100644 index 00000000000..d3fd00e75bf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/localVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='localVolumeSource', url='', help='Local represents directly-attached storage with node affinity (Beta feature)'), + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/main.libsonnet new file mode 100644 index 00000000000..44c09054bbc --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/main.libsonnet @@ -0,0 +1,175 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + affinity: (import 'affinity.libsonnet'), + attachedVolume: (import 'attachedVolume.libsonnet'), + awsElasticBlockStoreVolumeSource: (import 'awsElasticBlockStoreVolumeSource.libsonnet'), + azureDiskVolumeSource: (import 'azureDiskVolumeSource.libsonnet'), + azureFilePersistentVolumeSource: (import 'azureFilePersistentVolumeSource.libsonnet'), + azureFileVolumeSource: (import 'azureFileVolumeSource.libsonnet'), + binding: (import 'binding.libsonnet'), + capabilities: (import 'capabilities.libsonnet'), + cephFSPersistentVolumeSource: (import 'cephFSPersistentVolumeSource.libsonnet'), + cephFSVolumeSource: (import 'cephFSVolumeSource.libsonnet'), + cinderPersistentVolumeSource: (import 'cinderPersistentVolumeSource.libsonnet'), + cinderVolumeSource: (import 'cinderVolumeSource.libsonnet'), + clientIPConfig: (import 'clientIPConfig.libsonnet'), + componentCondition: (import 'componentCondition.libsonnet'), + componentStatus: (import 'componentStatus.libsonnet'), + configMap: (import 'configMap.libsonnet'), + configMapEnvSource: (import 'configMapEnvSource.libsonnet'), + configMapKeySelector: (import 'configMapKeySelector.libsonnet'), + configMapNodeConfigSource: (import 'configMapNodeConfigSource.libsonnet'), + configMapProjection: (import 'configMapProjection.libsonnet'), + configMapVolumeSource: (import 'configMapVolumeSource.libsonnet'), + container: (import 'container.libsonnet'), + containerImage: (import 'containerImage.libsonnet'), + containerPort: (import 'containerPort.libsonnet'), + containerState: (import 'containerState.libsonnet'), + containerStateRunning: (import 'containerStateRunning.libsonnet'), + containerStateTerminated: (import 'containerStateTerminated.libsonnet'), + containerStateWaiting: (import 'containerStateWaiting.libsonnet'), + containerStatus: (import 'containerStatus.libsonnet'), + csiPersistentVolumeSource: (import 'csiPersistentVolumeSource.libsonnet'), + csiVolumeSource: (import 'csiVolumeSource.libsonnet'), + daemonEndpoint: (import 'daemonEndpoint.libsonnet'), + downwardAPIProjection: (import 'downwardAPIProjection.libsonnet'), + downwardAPIVolumeFile: (import 'downwardAPIVolumeFile.libsonnet'), + downwardAPIVolumeSource: (import 'downwardAPIVolumeSource.libsonnet'), + emptyDirVolumeSource: (import 'emptyDirVolumeSource.libsonnet'), + endpointAddress: (import 'endpointAddress.libsonnet'), + endpointPort: (import 'endpointPort.libsonnet'), + endpointSubset: (import 'endpointSubset.libsonnet'), + endpoints: (import 'endpoints.libsonnet'), + envFromSource: (import 'envFromSource.libsonnet'), + envVar: (import 'envVar.libsonnet'), + envVarSource: (import 'envVarSource.libsonnet'), + ephemeralContainer: (import 'ephemeralContainer.libsonnet'), + ephemeralVolumeSource: (import 'ephemeralVolumeSource.libsonnet'), + event: (import 'event.libsonnet'), + eventSeries: (import 'eventSeries.libsonnet'), + eventSource: (import 'eventSource.libsonnet'), + execAction: (import 'execAction.libsonnet'), + fcVolumeSource: (import 'fcVolumeSource.libsonnet'), + flexPersistentVolumeSource: (import 'flexPersistentVolumeSource.libsonnet'), + flexVolumeSource: (import 'flexVolumeSource.libsonnet'), + flockerVolumeSource: (import 'flockerVolumeSource.libsonnet'), + gcePersistentDiskVolumeSource: (import 'gcePersistentDiskVolumeSource.libsonnet'), + gitRepoVolumeSource: (import 'gitRepoVolumeSource.libsonnet'), + glusterfsPersistentVolumeSource: (import 'glusterfsPersistentVolumeSource.libsonnet'), + glusterfsVolumeSource: (import 'glusterfsVolumeSource.libsonnet'), + handler: (import 'handler.libsonnet'), + hostAlias: (import 'hostAlias.libsonnet'), + hostPathVolumeSource: (import 'hostPathVolumeSource.libsonnet'), + httpGetAction: (import 'httpGetAction.libsonnet'), + httpHeader: (import 'httpHeader.libsonnet'), + iscsiPersistentVolumeSource: (import 'iscsiPersistentVolumeSource.libsonnet'), + iscsiVolumeSource: (import 'iscsiVolumeSource.libsonnet'), + keyToPath: (import 'keyToPath.libsonnet'), + lifecycle: (import 'lifecycle.libsonnet'), + limitRange: (import 'limitRange.libsonnet'), + limitRangeItem: (import 'limitRangeItem.libsonnet'), + limitRangeSpec: (import 'limitRangeSpec.libsonnet'), + loadBalancerIngress: (import 'loadBalancerIngress.libsonnet'), + loadBalancerStatus: (import 'loadBalancerStatus.libsonnet'), + localObjectReference: (import 'localObjectReference.libsonnet'), + localVolumeSource: (import 'localVolumeSource.libsonnet'), + namespace: (import 'namespace.libsonnet'), + namespaceCondition: (import 'namespaceCondition.libsonnet'), + namespaceSpec: (import 'namespaceSpec.libsonnet'), + namespaceStatus: (import 'namespaceStatus.libsonnet'), + nfsVolumeSource: (import 'nfsVolumeSource.libsonnet'), + node: (import 'node.libsonnet'), + nodeAddress: (import 'nodeAddress.libsonnet'), + nodeAffinity: (import 'nodeAffinity.libsonnet'), + nodeCondition: (import 'nodeCondition.libsonnet'), + nodeConfigSource: (import 'nodeConfigSource.libsonnet'), + nodeConfigStatus: (import 'nodeConfigStatus.libsonnet'), + nodeDaemonEndpoints: (import 'nodeDaemonEndpoints.libsonnet'), + nodeSelector: (import 'nodeSelector.libsonnet'), + nodeSelectorRequirement: (import 'nodeSelectorRequirement.libsonnet'), + nodeSelectorTerm: (import 'nodeSelectorTerm.libsonnet'), + nodeSpec: (import 'nodeSpec.libsonnet'), + nodeStatus: (import 'nodeStatus.libsonnet'), + nodeSystemInfo: (import 'nodeSystemInfo.libsonnet'), + objectFieldSelector: (import 'objectFieldSelector.libsonnet'), + objectReference: (import 'objectReference.libsonnet'), + persistentVolume: (import 'persistentVolume.libsonnet'), + persistentVolumeClaim: (import 'persistentVolumeClaim.libsonnet'), + persistentVolumeClaimCondition: (import 'persistentVolumeClaimCondition.libsonnet'), + persistentVolumeClaimSpec: (import 'persistentVolumeClaimSpec.libsonnet'), + persistentVolumeClaimStatus: (import 'persistentVolumeClaimStatus.libsonnet'), + persistentVolumeClaimTemplate: (import 'persistentVolumeClaimTemplate.libsonnet'), + persistentVolumeClaimVolumeSource: (import 'persistentVolumeClaimVolumeSource.libsonnet'), + persistentVolumeSpec: (import 'persistentVolumeSpec.libsonnet'), + persistentVolumeStatus: (import 'persistentVolumeStatus.libsonnet'), + photonPersistentDiskVolumeSource: (import 'photonPersistentDiskVolumeSource.libsonnet'), + pod: (import 'pod.libsonnet'), + podAffinity: (import 'podAffinity.libsonnet'), + podAffinityTerm: (import 'podAffinityTerm.libsonnet'), + podAntiAffinity: (import 'podAntiAffinity.libsonnet'), + podCondition: (import 'podCondition.libsonnet'), + podDNSConfig: (import 'podDNSConfig.libsonnet'), + podDNSConfigOption: (import 'podDNSConfigOption.libsonnet'), + podIP: (import 'podIP.libsonnet'), + podReadinessGate: (import 'podReadinessGate.libsonnet'), + podSecurityContext: (import 'podSecurityContext.libsonnet'), + podSpec: (import 'podSpec.libsonnet'), + podStatus: (import 'podStatus.libsonnet'), + podTemplate: (import 'podTemplate.libsonnet'), + podTemplateSpec: (import 'podTemplateSpec.libsonnet'), + portworxVolumeSource: (import 'portworxVolumeSource.libsonnet'), + preferredSchedulingTerm: (import 'preferredSchedulingTerm.libsonnet'), + probe: (import 'probe.libsonnet'), + projectedVolumeSource: (import 'projectedVolumeSource.libsonnet'), + quobyteVolumeSource: (import 'quobyteVolumeSource.libsonnet'), + rbdPersistentVolumeSource: (import 'rbdPersistentVolumeSource.libsonnet'), + rbdVolumeSource: (import 'rbdVolumeSource.libsonnet'), + replicationController: (import 'replicationController.libsonnet'), + replicationControllerCondition: (import 'replicationControllerCondition.libsonnet'), + replicationControllerSpec: (import 'replicationControllerSpec.libsonnet'), + replicationControllerStatus: (import 'replicationControllerStatus.libsonnet'), + resourceFieldSelector: (import 'resourceFieldSelector.libsonnet'), + resourceQuota: (import 'resourceQuota.libsonnet'), + resourceQuotaSpec: (import 'resourceQuotaSpec.libsonnet'), + resourceQuotaStatus: (import 'resourceQuotaStatus.libsonnet'), + resourceRequirements: (import 'resourceRequirements.libsonnet'), + scaleIOPersistentVolumeSource: (import 'scaleIOPersistentVolumeSource.libsonnet'), + scaleIOVolumeSource: (import 'scaleIOVolumeSource.libsonnet'), + scopeSelector: (import 'scopeSelector.libsonnet'), + scopedResourceSelectorRequirement: (import 'scopedResourceSelectorRequirement.libsonnet'), + seLinuxOptions: (import 'seLinuxOptions.libsonnet'), + seccompProfile: (import 'seccompProfile.libsonnet'), + secret: (import 'secret.libsonnet'), + secretEnvSource: (import 'secretEnvSource.libsonnet'), + secretKeySelector: (import 'secretKeySelector.libsonnet'), + secretProjection: (import 'secretProjection.libsonnet'), + secretReference: (import 'secretReference.libsonnet'), + secretVolumeSource: (import 'secretVolumeSource.libsonnet'), + securityContext: (import 'securityContext.libsonnet'), + service: (import 'service.libsonnet'), + serviceAccount: (import 'serviceAccount.libsonnet'), + serviceAccountTokenProjection: (import 'serviceAccountTokenProjection.libsonnet'), + servicePort: (import 'servicePort.libsonnet'), + serviceSpec: (import 'serviceSpec.libsonnet'), + serviceStatus: (import 'serviceStatus.libsonnet'), + sessionAffinityConfig: (import 'sessionAffinityConfig.libsonnet'), + storageOSPersistentVolumeSource: (import 'storageOSPersistentVolumeSource.libsonnet'), + storageOSVolumeSource: (import 'storageOSVolumeSource.libsonnet'), + sysctl: (import 'sysctl.libsonnet'), + taint: (import 'taint.libsonnet'), + tcpSocketAction: (import 'tcpSocketAction.libsonnet'), + toleration: (import 'toleration.libsonnet'), + topologySelectorLabelRequirement: (import 'topologySelectorLabelRequirement.libsonnet'), + topologySelectorTerm: (import 'topologySelectorTerm.libsonnet'), + topologySpreadConstraint: (import 'topologySpreadConstraint.libsonnet'), + typedLocalObjectReference: (import 'typedLocalObjectReference.libsonnet'), + volume: (import 'volume.libsonnet'), + volumeDevice: (import 'volumeDevice.libsonnet'), + volumeMount: (import 'volumeMount.libsonnet'), + volumeNodeAffinity: (import 'volumeNodeAffinity.libsonnet'), + volumeProjection: (import 'volumeProjection.libsonnet'), + vsphereVirtualDiskVolumeSource: (import 'vsphereVirtualDiskVolumeSource.libsonnet'), + weightedPodAffinityTerm: (import 'weightedPodAffinityTerm.libsonnet'), + windowsSecurityContextOptions: (import 'windowsSecurityContextOptions.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespace.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespace.libsonnet new file mode 100644 index 00000000000..da0733cd00d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespace.libsonnet @@ -0,0 +1,63 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='namespace', url='', help='Namespace provides a scope for Names. Use of multiple namespaces is optional.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Namespace', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Namespace', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='NamespaceSpec describes the attributes on a Namespace.'), + spec: { + '#withFinalizers':: d.fn(help='Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceCondition.libsonnet new file mode 100644 index 00000000000..14576467a92 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='namespaceCondition', url='', help='NamespaceCondition contains details about state of namespace.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help='', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of namespace controller condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceSpec.libsonnet new file mode 100644 index 00000000000..d95db4a1472 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceSpec.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='namespaceSpec', url='', help='NamespaceSpec describes the attributes on a Namespace.'), + '#withFinalizers':: d.fn(help='Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] }, + '#withFinalizersMixin':: d.fn(help='Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceStatus.libsonnet new file mode 100644 index 00000000000..f1657d85611 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/namespaceStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='namespaceStatus', url='', help='NamespaceStatus is information about the current status of a Namespace.'), + '#withConditions':: d.fn(help="Represents the latest available observations of a namespace's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a namespace's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withPhase':: d.fn(help='Phase is the current lifecycle phase of the namespace. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/', args=[d.arg(name='phase', type=d.T.string)]), + withPhase(phase): { phase: phase }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nfsVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nfsVolumeSource.libsonnet new file mode 100644 index 00000000000..cfdf14e1ce0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nfsVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nfsVolumeSource', url='', help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { server: server }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/node.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/node.libsonnet new file mode 100644 index 00000000000..cbde369c3fe --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/node.libsonnet @@ -0,0 +1,91 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='node', url='', help='Node is a worker node in Kubernetes. Each node will have a unique identifier in the cache (i.e. in etcd).'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Node', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Node', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='NodeSpec describes the attributes that a node is created with.'), + spec: { + '#configSource':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + configSource: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { spec+: { configSource+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { configSource+: { configMap+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { configSource+: { configMap+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { configSource+: { configMap+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { configSource+: { configMap+: { uid: uid } } } }, + }, + }, + '#withExternalID':: d.fn(help='Deprecated. Not all kubelets will set this field. Remove field after 1.13. see: https://issues.k8s.io/61966', args=[d.arg(name='externalID', type=d.T.string)]), + withExternalID(externalID): { spec+: { externalID: externalID } }, + '#withPodCIDR':: d.fn(help='PodCIDR represents the pod IP range assigned to the node.', args=[d.arg(name='podCIDR', type=d.T.string)]), + withPodCIDR(podCIDR): { spec+: { podCIDR: podCIDR } }, + '#withPodCIDRs':: d.fn(help='podCIDRs represents the IP ranges assigned to the node for usage by Pods on that node. If this field is specified, the 0th entry must match the podCIDR field. It may contain at most 1 value for each of IPv4 and IPv6.', args=[d.arg(name='podCIDRs', type=d.T.array)]), + withPodCIDRs(podCIDRs): { spec+: { podCIDRs: if std.isArray(v=podCIDRs) then podCIDRs else [podCIDRs] } }, + '#withPodCIDRsMixin':: d.fn(help='podCIDRs represents the IP ranges assigned to the node for usage by Pods on that node. If this field is specified, the 0th entry must match the podCIDR field. It may contain at most 1 value for each of IPv4 and IPv6.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podCIDRs', type=d.T.array)]), + withPodCIDRsMixin(podCIDRs): { spec+: { podCIDRs+: if std.isArray(v=podCIDRs) then podCIDRs else [podCIDRs] } }, + '#withProviderID':: d.fn(help='ID of the node assigned by the cloud provider in the format: ://', args=[d.arg(name='providerID', type=d.T.string)]), + withProviderID(providerID): { spec+: { providerID: providerID } }, + '#withTaints':: d.fn(help="If specified, the node's taints.", args=[d.arg(name='taints', type=d.T.array)]), + withTaints(taints): { spec+: { taints: if std.isArray(v=taints) then taints else [taints] } }, + '#withTaintsMixin':: d.fn(help="If specified, the node's taints.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='taints', type=d.T.array)]), + withTaintsMixin(taints): { spec+: { taints+: if std.isArray(v=taints) then taints else [taints] } }, + '#withUnschedulable':: d.fn(help='Unschedulable controls node schedulability of new pods. By default, node is schedulable. More info: https://kubernetes.io/docs/concepts/nodes/node/#manual-node-administration', args=[d.arg(name='unschedulable', type=d.T.boolean)]), + withUnschedulable(unschedulable): { spec+: { unschedulable: unschedulable } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAddress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAddress.libsonnet new file mode 100644 index 00000000000..cc6061a645d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAddress.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeAddress', url='', help="NodeAddress contains information for the node's address."), + '#withAddress':: d.fn(help='The node address.', args=[d.arg(name='address', type=d.T.string)]), + withAddress(address): { address: address }, + '#withType':: d.fn(help='Node address type, one of Hostname, ExternalIP or InternalIP.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAffinity.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAffinity.libsonnet new file mode 100644 index 00000000000..00c23f28ea1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeAffinity.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeAffinity', url='', help='Node affinity is a group of node affinity scheduling rules.'), + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeCondition.libsonnet new file mode 100644 index 00000000000..954593d911a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeCondition', url='', help='NodeCondition contains condition information for a node.'), + '#withLastHeartbeatTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastHeartbeatTime', type=d.T.string)]), + withLastHeartbeatTime(lastHeartbeatTime): { lastHeartbeatTime: lastHeartbeatTime }, + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="(brief) reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of node condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigSource.libsonnet new file mode 100644 index 00000000000..79954c7e7d1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigSource.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeConfigSource', url='', help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { configMap+: { kubeletConfigKey: kubeletConfigKey } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { configMap+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { configMap+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { configMap+: { uid: uid } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigStatus.libsonnet new file mode 100644 index 00000000000..338ed6227db --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeConfigStatus.libsonnet @@ -0,0 +1,56 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeConfigStatus', url='', help='NodeConfigStatus describes the status of the config assigned by Node.Spec.ConfigSource.'), + '#active':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + active: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { active+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { active+: { configMap+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { active+: { configMap+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { active+: { configMap+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { active+: { configMap+: { uid: uid } } }, + }, + }, + '#assigned':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + assigned: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { assigned+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { assigned+: { configMap+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { assigned+: { configMap+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { assigned+: { configMap+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { assigned+: { configMap+: { uid: uid } } }, + }, + }, + '#lastKnownGood':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + lastKnownGood: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { lastKnownGood+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { lastKnownGood+: { configMap+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { lastKnownGood+: { configMap+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { lastKnownGood+: { configMap+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { lastKnownGood+: { configMap+: { uid: uid } } }, + }, + }, + '#withError':: d.fn(help='Error describes any problems reconciling the Spec.ConfigSource to the Active config. Errors may occur, for example, attempting to checkpoint Spec.ConfigSource to the local Assigned record, attempting to checkpoint the payload associated with Spec.ConfigSource, attempting to load or validate the Assigned config, etc. Errors may occur at different points while syncing config. Earlier errors (e.g. download or checkpointing errors) will not result in a rollback to LastKnownGood, and may resolve across Kubelet retries. Later errors (e.g. loading or validating a checkpointed config) will result in a rollback to LastKnownGood. In the latter case, it is usually possible to resolve the error by fixing the config assigned in Spec.ConfigSource. You can find additional information for debugging by searching the error message in the Kubelet log. Error is a human-readable description of the error state; machines can check whether or not Error is empty, but should not rely on the stability of the Error text across Kubelet versions.', args=[d.arg(name='err', type=d.T.string)]), + withError(err): { 'error': err }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeDaemonEndpoints.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeDaemonEndpoints.libsonnet new file mode 100644 index 00000000000..c6579035537 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeDaemonEndpoints.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeDaemonEndpoints', url='', help='NodeDaemonEndpoints lists ports opened by daemons running on the Node.'), + '#kubeletEndpoint':: d.obj(help='DaemonEndpoint contains information about a single Daemon endpoint.'), + kubeletEndpoint: { + '#withPort':: d.fn(help='Port number of the given endpoint.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { kubeletEndpoint+: { port: port } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelector.libsonnet new file mode 100644 index 00000000000..ce1b27884ee --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelector.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeSelector', url='', help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorRequirement.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorRequirement.libsonnet new file mode 100644 index 00000000000..b039a232ff9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorRequirement.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeSelectorRequirement', url='', help='A node selector requirement is a selector that contains values, a key, and an operator that relates the key and values.'), + '#withKey':: d.fn(help='The label key that the selector applies to.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch.', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorTerm.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorTerm.libsonnet new file mode 100644 index 00000000000..4c4059da212 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSelectorTerm.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeSelectorTerm', url='', help='A null or empty node selector term matches no objects. The requirements of them are ANDed. The TopologySelectorTerm type implements a subset of the NodeSelectorTerm.'), + '#withMatchExpressions':: d.fn(help="A list of node selector requirements by node's labels.", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchExpressionsMixin':: d.fn(help="A list of node selector requirements by node's labels.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchFields':: d.fn(help="A list of node selector requirements by node's fields.", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + '#withMatchFieldsMixin':: d.fn(help="A list of node selector requirements by node's fields.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSpec.libsonnet new file mode 100644 index 00000000000..13cd0226bea --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSpec.libsonnet @@ -0,0 +1,38 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeSpec', url='', help='NodeSpec describes the attributes that a node is created with.'), + '#configSource':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + configSource: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { configSource+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configSource+: { configMap+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { configSource+: { configMap+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { configSource+: { configMap+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { configSource+: { configMap+: { uid: uid } } }, + }, + }, + '#withExternalID':: d.fn(help='Deprecated. Not all kubelets will set this field. Remove field after 1.13. see: https://issues.k8s.io/61966', args=[d.arg(name='externalID', type=d.T.string)]), + withExternalID(externalID): { externalID: externalID }, + '#withPodCIDR':: d.fn(help='PodCIDR represents the pod IP range assigned to the node.', args=[d.arg(name='podCIDR', type=d.T.string)]), + withPodCIDR(podCIDR): { podCIDR: podCIDR }, + '#withPodCIDRs':: d.fn(help='podCIDRs represents the IP ranges assigned to the node for usage by Pods on that node. If this field is specified, the 0th entry must match the podCIDR field. It may contain at most 1 value for each of IPv4 and IPv6.', args=[d.arg(name='podCIDRs', type=d.T.array)]), + withPodCIDRs(podCIDRs): { podCIDRs: if std.isArray(v=podCIDRs) then podCIDRs else [podCIDRs] }, + '#withPodCIDRsMixin':: d.fn(help='podCIDRs represents the IP ranges assigned to the node for usage by Pods on that node. If this field is specified, the 0th entry must match the podCIDR field. It may contain at most 1 value for each of IPv4 and IPv6.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podCIDRs', type=d.T.array)]), + withPodCIDRsMixin(podCIDRs): { podCIDRs+: if std.isArray(v=podCIDRs) then podCIDRs else [podCIDRs] }, + '#withProviderID':: d.fn(help='ID of the node assigned by the cloud provider in the format: ://', args=[d.arg(name='providerID', type=d.T.string)]), + withProviderID(providerID): { providerID: providerID }, + '#withTaints':: d.fn(help="If specified, the node's taints.", args=[d.arg(name='taints', type=d.T.array)]), + withTaints(taints): { taints: if std.isArray(v=taints) then taints else [taints] }, + '#withTaintsMixin':: d.fn(help="If specified, the node's taints.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='taints', type=d.T.array)]), + withTaintsMixin(taints): { taints+: if std.isArray(v=taints) then taints else [taints] }, + '#withUnschedulable':: d.fn(help='Unschedulable controls node schedulability of new pods. By default, node is schedulable. More info: https://kubernetes.io/docs/concepts/nodes/node/#manual-node-administration', args=[d.arg(name='unschedulable', type=d.T.boolean)]), + withUnschedulable(unschedulable): { unschedulable: unschedulable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeStatus.libsonnet new file mode 100644 index 00000000000..73cb8b044ad --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeStatus.libsonnet @@ -0,0 +1,120 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeStatus', url='', help='NodeStatus is information about the current status of a node.'), + '#config':: d.obj(help='NodeConfigStatus describes the status of the config assigned by Node.Spec.ConfigSource.'), + config: { + '#active':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + active: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { config+: { active+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { config+: { active+: { configMap+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { config+: { active+: { configMap+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { config+: { active+: { configMap+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { config+: { active+: { configMap+: { uid: uid } } } }, + }, + }, + '#assigned':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + assigned: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { config+: { assigned+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { config+: { assigned+: { configMap+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { config+: { assigned+: { configMap+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { config+: { assigned+: { configMap+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { config+: { assigned+: { configMap+: { uid: uid } } } }, + }, + }, + '#lastKnownGood':: d.obj(help='NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil.'), + lastKnownGood: { + '#configMap':: d.obj(help='ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node.'), + configMap: { + '#withKubeletConfigKey':: d.fn(help='KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases.', args=[d.arg(name='kubeletConfigKey', type=d.T.string)]), + withKubeletConfigKey(kubeletConfigKey): { config+: { lastKnownGood+: { configMap+: { kubeletConfigKey: kubeletConfigKey } } } }, + '#withName':: d.fn(help='Name is the metadata.name of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { config+: { lastKnownGood+: { configMap+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { config+: { lastKnownGood+: { configMap+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { config+: { lastKnownGood+: { configMap+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { config+: { lastKnownGood+: { configMap+: { uid: uid } } } }, + }, + }, + '#withError':: d.fn(help='Error describes any problems reconciling the Spec.ConfigSource to the Active config. Errors may occur, for example, attempting to checkpoint Spec.ConfigSource to the local Assigned record, attempting to checkpoint the payload associated with Spec.ConfigSource, attempting to load or validate the Assigned config, etc. Errors may occur at different points while syncing config. Earlier errors (e.g. download or checkpointing errors) will not result in a rollback to LastKnownGood, and may resolve across Kubelet retries. Later errors (e.g. loading or validating a checkpointed config) will result in a rollback to LastKnownGood. In the latter case, it is usually possible to resolve the error by fixing the config assigned in Spec.ConfigSource. You can find additional information for debugging by searching the error message in the Kubelet log. Error is a human-readable description of the error state; machines can check whether or not Error is empty, but should not rely on the stability of the Error text across Kubelet versions.', args=[d.arg(name='err', type=d.T.string)]), + withError(err): { config+: { 'error': err } }, + }, + '#daemonEndpoints':: d.obj(help='NodeDaemonEndpoints lists ports opened by daemons running on the Node.'), + daemonEndpoints: { + '#kubeletEndpoint':: d.obj(help='DaemonEndpoint contains information about a single Daemon endpoint.'), + kubeletEndpoint: { + '#withPort':: d.fn(help='Port number of the given endpoint.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { daemonEndpoints+: { kubeletEndpoint+: { port: port } } }, + }, + }, + '#nodeInfo':: d.obj(help='NodeSystemInfo is a set of ids/uuids to uniquely identify the node.'), + nodeInfo: { + '#withArchitecture':: d.fn(help='The Architecture reported by the node', args=[d.arg(name='architecture', type=d.T.string)]), + withArchitecture(architecture): { nodeInfo+: { architecture: architecture } }, + '#withBootID':: d.fn(help='Boot ID reported by the node.', args=[d.arg(name='bootID', type=d.T.string)]), + withBootID(bootID): { nodeInfo+: { bootID: bootID } }, + '#withContainerRuntimeVersion':: d.fn(help='ContainerRuntime Version reported by the node through runtime remote API (e.g. docker://1.5.0).', args=[d.arg(name='containerRuntimeVersion', type=d.T.string)]), + withContainerRuntimeVersion(containerRuntimeVersion): { nodeInfo+: { containerRuntimeVersion: containerRuntimeVersion } }, + '#withKernelVersion':: d.fn(help="Kernel Version reported by the node from 'uname -r' (e.g. 3.16.0-0.bpo.4-amd64).", args=[d.arg(name='kernelVersion', type=d.T.string)]), + withKernelVersion(kernelVersion): { nodeInfo+: { kernelVersion: kernelVersion } }, + '#withKubeProxyVersion':: d.fn(help='KubeProxy Version reported by the node.', args=[d.arg(name='kubeProxyVersion', type=d.T.string)]), + withKubeProxyVersion(kubeProxyVersion): { nodeInfo+: { kubeProxyVersion: kubeProxyVersion } }, + '#withKubeletVersion':: d.fn(help='Kubelet Version reported by the node.', args=[d.arg(name='kubeletVersion', type=d.T.string)]), + withKubeletVersion(kubeletVersion): { nodeInfo+: { kubeletVersion: kubeletVersion } }, + '#withMachineID':: d.fn(help='MachineID reported by the node. For unique machine identification in the cluster this field is preferred. Learn more from man(5) machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html', args=[d.arg(name='machineID', type=d.T.string)]), + withMachineID(machineID): { nodeInfo+: { machineID: machineID } }, + '#withOperatingSystem':: d.fn(help='The Operating System reported by the node', args=[d.arg(name='operatingSystem', type=d.T.string)]), + withOperatingSystem(operatingSystem): { nodeInfo+: { operatingSystem: operatingSystem } }, + '#withOsImage':: d.fn(help='OS Image reported by the node from /etc/os-release (e.g. Debian GNU/Linux 7 (wheezy)).', args=[d.arg(name='osImage', type=d.T.string)]), + withOsImage(osImage): { nodeInfo+: { osImage: osImage } }, + '#withSystemUUID':: d.fn(help='SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-us/red_hat_subscription_management/1/html/rhsm/uuid', args=[d.arg(name='systemUUID', type=d.T.string)]), + withSystemUUID(systemUUID): { nodeInfo+: { systemUUID: systemUUID } }, + }, + '#withAddresses':: d.fn(help='List of addresses reachable to the node. Queried from cloud provider, if available. More info: https://kubernetes.io/docs/concepts/nodes/node/#addresses Note: This field is declared as mergeable, but the merge key is not sufficiently unique, which can cause data corruption when it is merged. Callers should instead use a full-replacement patch. See http://pr.k8s.io/79391 for an example.', args=[d.arg(name='addresses', type=d.T.array)]), + withAddresses(addresses): { addresses: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withAddressesMixin':: d.fn(help='List of addresses reachable to the node. Queried from cloud provider, if available. More info: https://kubernetes.io/docs/concepts/nodes/node/#addresses Note: This field is declared as mergeable, but the merge key is not sufficiently unique, which can cause data corruption when it is merged. Callers should instead use a full-replacement patch. See http://pr.k8s.io/79391 for an example.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='addresses', type=d.T.array)]), + withAddressesMixin(addresses): { addresses+: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withAllocatable':: d.fn(help='Allocatable represents the resources of a node that are available for scheduling. Defaults to Capacity.', args=[d.arg(name='allocatable', type=d.T.object)]), + withAllocatable(allocatable): { allocatable: allocatable }, + '#withAllocatableMixin':: d.fn(help='Allocatable represents the resources of a node that are available for scheduling. Defaults to Capacity.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allocatable', type=d.T.object)]), + withAllocatableMixin(allocatable): { allocatable+: allocatable }, + '#withCapacity':: d.fn(help='Capacity represents the total resources of a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity', args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { capacity: capacity }, + '#withCapacityMixin':: d.fn(help='Capacity represents the total resources of a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { capacity+: capacity }, + '#withConditions':: d.fn(help='Conditions is an array of current observed node conditions. More info: https://kubernetes.io/docs/concepts/nodes/node/#condition', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='Conditions is an array of current observed node conditions. More info: https://kubernetes.io/docs/concepts/nodes/node/#condition\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withImages':: d.fn(help='List of container images on this node', args=[d.arg(name='images', type=d.T.array)]), + withImages(images): { images: if std.isArray(v=images) then images else [images] }, + '#withImagesMixin':: d.fn(help='List of container images on this node\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='images', type=d.T.array)]), + withImagesMixin(images): { images+: if std.isArray(v=images) then images else [images] }, + '#withPhase':: d.fn(help='NodePhase is the recently observed lifecycle phase of the node. More info: https://kubernetes.io/docs/concepts/nodes/node/#phase The field is never populated, and now is deprecated.', args=[d.arg(name='phase', type=d.T.string)]), + withPhase(phase): { phase: phase }, + '#withVolumesAttached':: d.fn(help='List of volumes that are attached to the node.', args=[d.arg(name='volumesAttached', type=d.T.array)]), + withVolumesAttached(volumesAttached): { volumesAttached: if std.isArray(v=volumesAttached) then volumesAttached else [volumesAttached] }, + '#withVolumesAttachedMixin':: d.fn(help='List of volumes that are attached to the node.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumesAttached', type=d.T.array)]), + withVolumesAttachedMixin(volumesAttached): { volumesAttached+: if std.isArray(v=volumesAttached) then volumesAttached else [volumesAttached] }, + '#withVolumesInUse':: d.fn(help='List of attachable volumes in use (mounted) by the node.', args=[d.arg(name='volumesInUse', type=d.T.array)]), + withVolumesInUse(volumesInUse): { volumesInUse: if std.isArray(v=volumesInUse) then volumesInUse else [volumesInUse] }, + '#withVolumesInUseMixin':: d.fn(help='List of attachable volumes in use (mounted) by the node.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumesInUse', type=d.T.array)]), + withVolumesInUseMixin(volumesInUse): { volumesInUse+: if std.isArray(v=volumesInUse) then volumesInUse else [volumesInUse] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSystemInfo.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSystemInfo.libsonnet new file mode 100644 index 00000000000..35eff51ffa5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/nodeSystemInfo.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nodeSystemInfo', url='', help='NodeSystemInfo is a set of ids/uuids to uniquely identify the node.'), + '#withArchitecture':: d.fn(help='The Architecture reported by the node', args=[d.arg(name='architecture', type=d.T.string)]), + withArchitecture(architecture): { architecture: architecture }, + '#withBootID':: d.fn(help='Boot ID reported by the node.', args=[d.arg(name='bootID', type=d.T.string)]), + withBootID(bootID): { bootID: bootID }, + '#withContainerRuntimeVersion':: d.fn(help='ContainerRuntime Version reported by the node through runtime remote API (e.g. docker://1.5.0).', args=[d.arg(name='containerRuntimeVersion', type=d.T.string)]), + withContainerRuntimeVersion(containerRuntimeVersion): { containerRuntimeVersion: containerRuntimeVersion }, + '#withKernelVersion':: d.fn(help="Kernel Version reported by the node from 'uname -r' (e.g. 3.16.0-0.bpo.4-amd64).", args=[d.arg(name='kernelVersion', type=d.T.string)]), + withKernelVersion(kernelVersion): { kernelVersion: kernelVersion }, + '#withKubeProxyVersion':: d.fn(help='KubeProxy Version reported by the node.', args=[d.arg(name='kubeProxyVersion', type=d.T.string)]), + withKubeProxyVersion(kubeProxyVersion): { kubeProxyVersion: kubeProxyVersion }, + '#withKubeletVersion':: d.fn(help='Kubelet Version reported by the node.', args=[d.arg(name='kubeletVersion', type=d.T.string)]), + withKubeletVersion(kubeletVersion): { kubeletVersion: kubeletVersion }, + '#withMachineID':: d.fn(help='MachineID reported by the node. For unique machine identification in the cluster this field is preferred. Learn more from man(5) machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html', args=[d.arg(name='machineID', type=d.T.string)]), + withMachineID(machineID): { machineID: machineID }, + '#withOperatingSystem':: d.fn(help='The Operating System reported by the node', args=[d.arg(name='operatingSystem', type=d.T.string)]), + withOperatingSystem(operatingSystem): { operatingSystem: operatingSystem }, + '#withOsImage':: d.fn(help='OS Image reported by the node from /etc/os-release (e.g. Debian GNU/Linux 7 (wheezy)).', args=[d.arg(name='osImage', type=d.T.string)]), + withOsImage(osImage): { osImage: osImage }, + '#withSystemUUID':: d.fn(help='SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-us/red_hat_subscription_management/1/html/rhsm/uuid', args=[d.arg(name='systemUUID', type=d.T.string)]), + withSystemUUID(systemUUID): { systemUUID: systemUUID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectFieldSelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectFieldSelector.libsonnet new file mode 100644 index 00000000000..a65f4a01db9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectFieldSelector.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectFieldSelector', url='', help='ObjectFieldSelector selects an APIVersioned field of an object.'), + '#withFieldPath':: d.fn(help='Path of the field to select in the specified API version.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldPath: fieldPath }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectReference.libsonnet new file mode 100644 index 00000000000..9f40df55596 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/objectReference.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='objectReference', url='', help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldPath: fieldPath }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { resourceVersion: resourceVersion }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { uid: uid }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolume.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolume.libsonnet new file mode 100644 index 00000000000..a12fc70180a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolume.libsonnet @@ -0,0 +1,467 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolume', url='', help='PersistentVolume (PV) is a storage resource provisioned by an administrator. It is analogous to a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Persistentvolume', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'PersistentVolume', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + spec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { awsElasticBlockStore+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { awsElasticBlockStore+: { partition: partition } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { awsElasticBlockStore+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { awsElasticBlockStore+: { volumeID: volumeID } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { spec+: { azureDisk+: { cachingMode: cachingMode } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { spec+: { azureDisk+: { diskName: diskName } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { spec+: { azureDisk+: { diskURI: diskURI } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { azureDisk+: { fsType: fsType } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { azureDisk+: { kind: kind } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { azureDisk+: { readOnly: readOnly } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { azureFile+: { readOnly: readOnly } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { spec+: { azureFile+: { secretName: secretName } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { spec+: { azureFile+: { secretNamespace: secretNamespace } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { spec+: { azureFile+: { shareName: shareName } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { cephfs+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { cephfs+: { secretRef+: { namespace: namespace } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { cephfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { cephfs+: { readOnly: readOnly } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { spec+: { cephfs+: { secretFile: secretFile } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { cephfs+: { user: user } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { cinder+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { cinder+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { cinder+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { cinder+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { cinder+: { volumeID: volumeID } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { claimRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { claimRef+: { fieldPath: fieldPath } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { claimRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { claimRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { claimRef+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { claimRef+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { claimRef+: { uid: uid } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { csi+: { controllerExpandSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { csi+: { controllerPublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { csi+: { nodePublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { csi+: { nodeStageSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { csi+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { csi+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { csi+: { readOnly: readOnly } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { spec+: { csi+: { volumeAttributes: volumeAttributes } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { spec+: { csi+: { volumeAttributes+: volumeAttributes } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { spec+: { csi+: { volumeHandle: volumeHandle } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { fc+: { fsType: fsType } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { fc+: { lun: lun } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { fc+: { readOnly: readOnly } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { spec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { spec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { spec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { spec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { flexVolume+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { flexVolume+: { secretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { flexVolume+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { flexVolume+: { fsType: fsType } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { spec+: { flexVolume+: { options: options } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { spec+: { flexVolume+: { options+: options } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { flexVolume+: { readOnly: readOnly } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { spec+: { flocker+: { datasetName: datasetName } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { spec+: { flocker+: { datasetUUID: datasetUUID } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { gcePersistentDisk+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { gcePersistentDisk+: { partition: partition } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { spec+: { gcePersistentDisk+: { pdName: pdName } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { gcePersistentDisk+: { readOnly: readOnly } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { spec+: { glusterfs+: { endpoints: endpoints } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { spec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { glusterfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { glusterfs+: { readOnly: readOnly } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { hostPath+: { path: path } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { hostPath+: { type: type } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { iscsi+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { iscsi+: { secretRef+: { namespace: namespace } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { spec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { spec+: { iscsi+: { chapAuthSession: chapAuthSession } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { iscsi+: { fsType: fsType } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { spec+: { iscsi+: { initiatorName: initiatorName } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { spec+: { iscsi+: { iqn: iqn } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { spec+: { iscsi+: { iscsiInterface: iscsiInterface } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { iscsi+: { lun: lun } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { spec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { spec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { iscsi+: { readOnly: readOnly } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { spec+: { iscsi+: { targetPortal: targetPortal } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { 'local'+: { fsType: fsType } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { 'local'+: { path: path } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { nfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { nfs+: { readOnly: readOnly } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { spec+: { nfs+: { server: server } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { photonPersistentDisk+: { fsType: fsType } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { spec+: { photonPersistentDisk+: { pdID: pdID } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { portworxVolume+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { portworxVolume+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { portworxVolume+: { volumeID: volumeID } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { quobyte+: { group: group } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { quobyte+: { readOnly: readOnly } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { spec+: { quobyte+: { registry: registry } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { spec+: { quobyte+: { tenant: tenant } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { quobyte+: { user: user } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { spec+: { quobyte+: { volume: volume } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { rbd+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { rbd+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { rbd+: { fsType: fsType } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { rbd+: { image: image } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { spec+: { rbd+: { keyring: keyring } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { spec+: { rbd+: { pool: pool } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { rbd+: { readOnly: readOnly } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { rbd+: { user: user } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { scaleIO+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { scaleIO+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { scaleIO+: { fsType: fsType } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { spec+: { scaleIO+: { gateway: gateway } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { spec+: { scaleIO+: { protectionDomain: protectionDomain } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { scaleIO+: { readOnly: readOnly } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { spec+: { scaleIO+: { sslEnabled: sslEnabled } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { spec+: { scaleIO+: { storageMode: storageMode } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { spec+: { scaleIO+: { storagePool: storagePool } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { spec+: { scaleIO+: { system: system } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { scaleIO+: { volumeName: volumeName } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storageos+: { secretRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storageos+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { storageos+: { secretRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { storageos+: { secretRef+: { uid: uid } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { storageos+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { storageos+: { readOnly: readOnly } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { storageos+: { volumeName: volumeName } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { spec+: { storageos+: { volumeNamespace: volumeNamespace } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { vsphereVolume+: { fsType: fsType } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { spec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { spec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { spec+: { vsphereVolume+: { volumePath: volumePath } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { spec+: { capacity: capacity } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { spec+: { capacity+: capacity } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { spec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { spec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { spec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { volumeMode: volumeMode } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaim.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaim.libsonnet new file mode 100644 index 00000000000..26c44dcc4af --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaim.libsonnet @@ -0,0 +1,100 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaim', url='', help="PersistentVolumeClaim is a user's request for and claim to a persistent volume"), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Persistentvolumeclaim', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'PersistentVolumeClaim', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes'), + spec: { + '#dataSource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + dataSource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { dataSource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { dataSource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { dataSource+: { name: name } } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { resources+: { limits: limits } } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { resources+: { limits+: limits } } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { resources+: { requests: requests } } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { resources+: { requests+: requests } } }, + }, + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withStorageClassName':: d.fn(help='Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { volumeMode: volumeMode } }, + '#withVolumeName':: d.fn(help='VolumeName is the binding reference to the PersistentVolume backing this claim.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { volumeName: volumeName } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimCondition.libsonnet new file mode 100644 index 00000000000..8773c355aac --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaimCondition', url='', help='PersistentVolumeClaimCondition contails details about state of pvc'), + '#withLastProbeTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastProbeTime', type=d.T.string)]), + withLastProbeTime(lastProbeTime): { lastProbeTime: lastProbeTime }, + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="Unique, this should be a short, machine understandable string that gives the reason for condition's last transition. If it reports 'ResizeStarted' that means the underlying persistent volume is being resized.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimSpec.libsonnet new file mode 100644 index 00000000000..22091ea4ad3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimSpec.libsonnet @@ -0,0 +1,47 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaimSpec', url='', help='PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes'), + '#dataSource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + dataSource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { dataSource+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { dataSource+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { dataSource+: { name: name } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withStorageClassName':: d.fn(help='Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { storageClassName: storageClassName }, + '#withVolumeMode':: d.fn(help='volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { volumeMode: volumeMode }, + '#withVolumeName':: d.fn(help='VolumeName is the binding reference to the PersistentVolume backing this claim.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeName: volumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimStatus.libsonnet new file mode 100644 index 00000000000..4053907baea --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimStatus.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaimStatus', url='', help='PersistentVolumeClaimStatus is the current status of a persistent volume claim.'), + '#withAccessModes':: d.fn(help='AccessModes contains the actual access modes the volume backing the PVC has. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the actual access modes the volume backing the PVC has. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withCapacity':: d.fn(help='Represents the actual resources of the underlying volume.', args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { capacity: capacity }, + '#withCapacityMixin':: d.fn(help='Represents the actual resources of the underlying volume.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { capacity+: capacity }, + '#withConditions':: d.fn(help="Current Condition of persistent volume claim. If underlying persistent volume is being resized then the Condition will be set to 'ResizeStarted'.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Current Condition of persistent volume claim. If underlying persistent volume is being resized then the Condition will be set to 'ResizeStarted'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withPhase':: d.fn(help='Phase represents the current phase of PersistentVolumeClaim.', args=[d.arg(name='phase', type=d.T.string)]), + withPhase(phase): { phase: phase }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimTemplate.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimTemplate.libsonnet new file mode 100644 index 00000000000..7760c3cd422 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimTemplate.libsonnet @@ -0,0 +1,95 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaimTemplate', url='', help='PersistentVolumeClaimTemplate is used to produce PersistentVolumeClaim objects as part of an EphemeralVolumeSource.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#spec':: d.obj(help='PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes'), + spec: { + '#dataSource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + dataSource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { dataSource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { dataSource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { dataSource+: { name: name } } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { resources+: { limits: limits } } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { resources+: { limits+: limits } } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { resources+: { requests: requests } } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { resources+: { requests+: requests } } }, + }, + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withStorageClassName':: d.fn(help='Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { volumeMode: volumeMode } }, + '#withVolumeName':: d.fn(help='VolumeName is the binding reference to the PersistentVolume backing this claim.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { volumeName: volumeName } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimVolumeSource.libsonnet new file mode 100644 index 00000000000..6a30b0d4b9b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeClaimVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeClaimVolumeSource', url='', help="PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. This volume finds the bound PV and mounts that volume for the pod. A PersistentVolumeClaimVolumeSource is, essentially, a wrapper around another type of volume that is owned by someone else (the system)."), + '#withClaimName':: d.fn(help='ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims', args=[d.arg(name='claimName', type=d.T.string)]), + withClaimName(claimName): { claimName: claimName }, + '#withReadOnly':: d.fn(help='Will force the ReadOnly setting in VolumeMounts. Default false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeSpec.libsonnet new file mode 100644 index 00000000000..74c2a7d7252 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeSpec.libsonnet @@ -0,0 +1,414 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeSpec', url='', help='PersistentVolumeSpec is the specification of a persistent volume.'), + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { awsElasticBlockStore+: { fsType: fsType } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { awsElasticBlockStore+: { partition: partition } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { awsElasticBlockStore+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { awsElasticBlockStore+: { volumeID: volumeID } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { azureDisk+: { cachingMode: cachingMode } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { azureDisk+: { diskName: diskName } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { azureDisk+: { diskURI: diskURI } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { azureDisk+: { fsType: fsType } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { azureDisk+: { kind: kind } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureDisk+: { readOnly: readOnly } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureFile+: { readOnly: readOnly } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { azureFile+: { secretName: secretName } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { azureFile+: { secretNamespace: secretNamespace } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { azureFile+: { shareName: shareName } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cephfs+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { cephfs+: { secretRef+: { namespace: namespace } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { cephfs+: { path: path } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cephfs+: { readOnly: readOnly } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { cephfs+: { secretFile: secretFile } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { cephfs+: { user: user } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cinder+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { cinder+: { secretRef+: { namespace: namespace } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { cinder+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cinder+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { cinder+: { volumeID: volumeID } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { claimRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { claimRef+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { claimRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { claimRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { claimRef+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { claimRef+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { claimRef+: { uid: uid } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { controllerExpandSecretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { csi+: { controllerExpandSecretRef+: { namespace: namespace } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { controllerPublishSecretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { csi+: { controllerPublishSecretRef+: { namespace: namespace } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { nodePublishSecretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { csi+: { nodePublishSecretRef+: { namespace: namespace } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { nodeStageSecretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { csi+: { nodeStageSecretRef+: { namespace: namespace } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { csi+: { driver: driver } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { csi+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { csi+: { readOnly: readOnly } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { csi+: { volumeAttributes: volumeAttributes } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { csi+: { volumeAttributes+: volumeAttributes } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { csi+: { volumeHandle: volumeHandle } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fc+: { fsType: fsType } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { fc+: { lun: lun } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { fc+: { readOnly: readOnly } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { flexVolume+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { flexVolume+: { secretRef+: { namespace: namespace } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { flexVolume+: { driver: driver } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { flexVolume+: { fsType: fsType } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { flexVolume+: { options: options } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { flexVolume+: { options+: options } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { flexVolume+: { readOnly: readOnly } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { flocker+: { datasetName: datasetName } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { flocker+: { datasetUUID: datasetUUID } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { gcePersistentDisk+: { fsType: fsType } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { gcePersistentDisk+: { partition: partition } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { gcePersistentDisk+: { pdName: pdName } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { gcePersistentDisk+: { readOnly: readOnly } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { glusterfs+: { endpoints: endpoints } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { glusterfs+: { endpointsNamespace: endpointsNamespace } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { glusterfs+: { path: path } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { glusterfs+: { readOnly: readOnly } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { hostPath+: { path: path } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { hostPath+: { type: type } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { iscsi+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { iscsi+: { secretRef+: { namespace: namespace } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { iscsi+: { chapAuthSession: chapAuthSession } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { iscsi+: { fsType: fsType } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { iscsi+: { initiatorName: initiatorName } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iscsi+: { iqn: iqn } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsi+: { iscsiInterface: iscsiInterface } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { iscsi+: { lun: lun } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { iscsi+: { readOnly: readOnly } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { iscsi+: { targetPortal: targetPortal } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { 'local'+: { fsType: fsType } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { 'local'+: { path: path } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nfs+: { path: path } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { nfs+: { readOnly: readOnly } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { nfs+: { server: server } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { photonPersistentDisk+: { fsType: fsType } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { photonPersistentDisk+: { pdID: pdID } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { portworxVolume+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { portworxVolume+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { portworxVolume+: { volumeID: volumeID } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { quobyte+: { group: group } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { quobyte+: { readOnly: readOnly } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { quobyte+: { registry: registry } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { quobyte+: { tenant: tenant } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { quobyte+: { user: user } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { quobyte+: { volume: volume } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { rbd+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { rbd+: { secretRef+: { namespace: namespace } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { rbd+: { fsType: fsType } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { rbd+: { image: image } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { rbd+: { keyring: keyring } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { rbd+: { pool: pool } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { rbd+: { readOnly: readOnly } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { rbd+: { user: user } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleIO+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { scaleIO+: { secretRef+: { namespace: namespace } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { scaleIO+: { fsType: fsType } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { scaleIO+: { gateway: gateway } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { scaleIO+: { protectionDomain: protectionDomain } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { scaleIO+: { readOnly: readOnly } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { scaleIO+: { sslEnabled: sslEnabled } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { scaleIO+: { storageMode: storageMode } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { scaleIO+: { storagePool: storagePool } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { scaleIO+: { system: system } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { scaleIO+: { volumeName: volumeName } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { storageos+: { secretRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { storageos+: { secretRef+: { fieldPath: fieldPath } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { storageos+: { secretRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { storageos+: { secretRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { storageos+: { secretRef+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { storageos+: { secretRef+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { storageos+: { secretRef+: { uid: uid } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { storageos+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { storageos+: { readOnly: readOnly } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { storageos+: { volumeName: volumeName } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { storageos+: { volumeNamespace: volumeNamespace } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { vsphereVolume+: { fsType: fsType } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { vsphereVolume+: { storagePolicyID: storagePolicyID } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { vsphereVolume+: { storagePolicyName: storagePolicyName } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { vsphereVolume+: { volumePath: volumePath } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { capacity: capacity }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { capacity+: capacity }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { storageClassName: storageClassName }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { volumeMode: volumeMode }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeStatus.libsonnet new file mode 100644 index 00000000000..dde767c2d18 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/persistentVolumeStatus.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='persistentVolumeStatus', url='', help='PersistentVolumeStatus is the current status of a persistent volume.'), + '#withMessage':: d.fn(help='A human-readable message indicating details about why the volume is in this state.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withPhase':: d.fn(help='Phase indicates if a volume is available, bound to a claim, or released by a claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#phase', args=[d.arg(name='phase', type=d.T.string)]), + withPhase(phase): { phase: phase }, + '#withReason':: d.fn(help='Reason is a brief CamelCase string that describes any failure and is meant for machine parsing and tidy display in the CLI.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/photonPersistentDiskVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/photonPersistentDiskVolumeSource.libsonnet new file mode 100644 index 00000000000..d8c1209ede9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/photonPersistentDiskVolumeSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='photonPersistentDiskVolumeSource', url='', help='Represents a Photon Controller persistent disk resource.'), + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { pdID: pdID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/pod.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/pod.libsonnet new file mode 100644 index 00000000000..0862ad510eb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/pod.libsonnet @@ -0,0 +1,247 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='pod', url='', help='Pod is a collection of containers that can run on a host. This resource is created by clients and scheduled onto hosts.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Pod', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Pod', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seccompProfile+: { type: type } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { securityContext+: { fsGroup: fsGroup } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { securityContext+: { runAsGroup: runAsGroup } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { securityContext+: { runAsUser: runAsUser } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { automountServiceAccountToken: automountServiceAccountToken } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { dnsPolicy: dnsPolicy } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { enableServiceLinks: enableServiceLinks } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { hostIPC: hostIPC } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { hostNetwork: hostNetwork } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { hostPID: hostPID } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { hostname: hostname } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { nodeName: nodeName } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { nodeSelector+: nodeSelector } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { overhead: overhead } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { overhead+: overhead } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { preemptionPolicy: preemptionPolicy } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { priority: priority } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { restartPolicy: restartPolicy } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { runtimeClassName: runtimeClassName } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { serviceAccount: serviceAccount } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { serviceAccountName: serviceAccountName } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { shareProcessNamespace: shareProcessNamespace } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { subdomain: subdomain } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinity.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinity.libsonnet new file mode 100644 index 00000000000..18ac76821b3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinity.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podAffinity', url='', help='Pod affinity is a group of inter pod affinity scheduling rules.'), + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinityTerm.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinityTerm.libsonnet new file mode 100644 index 00000000000..cb3106e2874 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAffinityTerm.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podAffinityTerm', url='', help='Defines a set of pods (namely those matching the labelSelector relative to the given namespace(s)) that this pod should be co-located (affinity) or not co-located (anti-affinity) with, where co-located is defined as running on a node whose value of the label with key matches that of any node on which a pod of the set of pods is running'), + '#labelSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + labelSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#withNamespaces':: d.fn(help="namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means 'this pod's namespace'", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help="namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means 'this pod's namespace'\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withTopologyKey':: d.fn(help='This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed.', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAntiAffinity.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAntiAffinity.libsonnet new file mode 100644 index 00000000000..55a58d3aef9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podAntiAffinity.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podAntiAffinity', url='', help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podCondition.libsonnet new file mode 100644 index 00000000000..f3e7febbebb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podCondition.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podCondition', url='', help='PodCondition contains details for the current condition of this pod.'), + '#withLastProbeTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastProbeTime', type=d.T.string)]), + withLastProbeTime(lastProbeTime): { lastProbeTime: lastProbeTime }, + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='Human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="Unique, one-word, CamelCase reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type is the type of the condition. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfig.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfig.libsonnet new file mode 100644 index 00000000000..531e273b1ce --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfig.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podDNSConfig', url='', help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { options: if std.isArray(v=options) then options else [options] }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { options+: if std.isArray(v=options) then options else [options] }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { searches: if std.isArray(v=searches) then searches else [searches] }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { searches+: if std.isArray(v=searches) then searches else [searches] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfigOption.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfigOption.libsonnet new file mode 100644 index 00000000000..e346886097a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podDNSConfigOption.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podDNSConfigOption', url='', help='PodDNSConfigOption defines DNS resolver options of a pod.'), + '#withName':: d.fn(help='Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podIP.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podIP.libsonnet new file mode 100644 index 00000000000..3624968ef4e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podIP.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podIP', url='', help='IP address information for entries in the (plural) PodIPs field. Each entry includes:\n IP: An IP address allocated to the pod. Routable at least within the cluster.'), + '#withIp':: d.fn(help='ip is an IP address (IPv4 or IPv6) assigned to the pod', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podReadinessGate.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podReadinessGate.libsonnet new file mode 100644 index 00000000000..e0411927ec9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podReadinessGate.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podReadinessGate', url='', help='PodReadinessGate contains the reference to a pod condition'), + '#withConditionType':: d.fn(help="ConditionType refers to a condition in the pod's condition list with matching type.", args=[d.arg(name='conditionType', type=d.T.string)]), + withConditionType(conditionType): { conditionType: conditionType }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSecurityContext.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSecurityContext.libsonnet new file mode 100644 index 00000000000..e792ea53a62 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSecurityContext.libsonnet @@ -0,0 +1,51 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podSecurityContext', url='', help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { seLinuxOptions+: { level: level } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { seLinuxOptions+: { role: role } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seLinuxOptions+: { type: type } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { seLinuxOptions+: { user: user } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { seccompProfile+: { localhostProfile: localhostProfile } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seccompProfile+: { type: type } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { windowsOptions+: { runAsUserName: runAsUserName } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { fsGroup: fsGroup }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { fsGroupChangePolicy: fsGroupChangePolicy }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { runAsGroup: runAsGroup }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { runAsNonRoot: runAsNonRoot }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { runAsUser: runAsUser }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSpec.libsonnet new file mode 100644 index 00000000000..8158694ab14 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podSpec.libsonnet @@ -0,0 +1,194 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podSpec', url='', help='PodSpec is a description of a pod.'), + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { securityContext+: { fsGroup: fsGroup } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { activeDeadlineSeconds: activeDeadlineSeconds }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { automountServiceAccountToken: automountServiceAccountToken }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { containers: if std.isArray(v=containers) then containers else [containers] }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { containers+: if std.isArray(v=containers) then containers else [containers] }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { dnsPolicy: dnsPolicy }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { enableServiceLinks: enableServiceLinks }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { hostIPC: hostIPC }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { hostNetwork: hostNetwork }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { hostPID: hostPID }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { hostname: hostname }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { nodeName: nodeName }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { nodeSelector: nodeSelector }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { nodeSelector+: nodeSelector }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { overhead: overhead }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { overhead+: overhead }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { preemptionPolicy: preemptionPolicy }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { priority: priority }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { priorityClassName: priorityClassName }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { runtimeClassName: runtimeClassName }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { schedulerName: schedulerName }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { serviceAccount: serviceAccount }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { serviceAccountName: serviceAccountName }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { setHostnameAsFQDN: setHostnameAsFQDN }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { shareProcessNamespace: shareProcessNamespace }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { subdomain: subdomain }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { terminationGracePeriodSeconds: terminationGracePeriodSeconds }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { volumes: if std.isArray(v=volumes) then volumes else [volumes] }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { volumes+: if std.isArray(v=volumes) then volumes else [volumes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podStatus.libsonnet new file mode 100644 index 00000000000..8c3b8f1f8f6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podStatus.libsonnet @@ -0,0 +1,42 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podStatus', url='', help='PodStatus represents information about the status of a pod. Status may trail the actual state of a system, especially if the node that hosts the pod cannot contact the control plane.'), + '#withConditions':: d.fn(help='Current service state of pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='Current service state of pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withContainerStatuses':: d.fn(help='The list has one entry per container in the manifest. Each entry is currently the output of `docker inspect`. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status', args=[d.arg(name='containerStatuses', type=d.T.array)]), + withContainerStatuses(containerStatuses): { containerStatuses: if std.isArray(v=containerStatuses) then containerStatuses else [containerStatuses] }, + '#withContainerStatusesMixin':: d.fn(help='The list has one entry per container in the manifest. Each entry is currently the output of `docker inspect`. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containerStatuses', type=d.T.array)]), + withContainerStatusesMixin(containerStatuses): { containerStatuses+: if std.isArray(v=containerStatuses) then containerStatuses else [containerStatuses] }, + '#withEphemeralContainerStatuses':: d.fn(help='Status for any ephemeral containers that have run in this pod. This field is alpha-level and is only populated by servers that enable the EphemeralContainers feature.', args=[d.arg(name='ephemeralContainerStatuses', type=d.T.array)]), + withEphemeralContainerStatuses(ephemeralContainerStatuses): { ephemeralContainerStatuses: if std.isArray(v=ephemeralContainerStatuses) then ephemeralContainerStatuses else [ephemeralContainerStatuses] }, + '#withEphemeralContainerStatusesMixin':: d.fn(help='Status for any ephemeral containers that have run in this pod. This field is alpha-level and is only populated by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ephemeralContainerStatuses', type=d.T.array)]), + withEphemeralContainerStatusesMixin(ephemeralContainerStatuses): { ephemeralContainerStatuses+: if std.isArray(v=ephemeralContainerStatuses) then ephemeralContainerStatuses else [ephemeralContainerStatuses] }, + '#withHostIP':: d.fn(help='IP address of the host to which the pod is assigned. Empty if not yet scheduled.', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withInitContainerStatuses':: d.fn(help='The list has one entry per init container in the manifest. The most recent successful init container will have ready = true, the most recently started container will have startTime set. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status', args=[d.arg(name='initContainerStatuses', type=d.T.array)]), + withInitContainerStatuses(initContainerStatuses): { initContainerStatuses: if std.isArray(v=initContainerStatuses) then initContainerStatuses else [initContainerStatuses] }, + '#withInitContainerStatusesMixin':: d.fn(help='The list has one entry per init container in the manifest. The most recent successful init container will have ready = true, the most recently started container will have startTime set. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainerStatuses', type=d.T.array)]), + withInitContainerStatusesMixin(initContainerStatuses): { initContainerStatuses+: if std.isArray(v=initContainerStatuses) then initContainerStatuses else [initContainerStatuses] }, + '#withMessage':: d.fn(help='A human readable message indicating details about why the pod is in this condition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withNominatedNodeName':: d.fn(help='nominatedNodeName is set only when this pod preempts other pods on the node, but it cannot be scheduled right away as preemption victims receive their graceful termination periods. This field does not guarantee that the pod will be scheduled on this node. Scheduler may decide to place the pod elsewhere if other nodes become available sooner. Scheduler may also decide to give the resources on this node to a higher priority pod that is created after preemption. As a result, this field may be different than PodSpec.nodeName when the pod is scheduled.', args=[d.arg(name='nominatedNodeName', type=d.T.string)]), + withNominatedNodeName(nominatedNodeName): { nominatedNodeName: nominatedNodeName }, + '#withPhase':: d.fn(help="The phase of a Pod is a simple, high-level summary of where the Pod is in its lifecycle. The conditions array, the reason and message fields, and the individual container status arrays contain more detail about the pod's status. There are five possible phase values:\n\nPending: The pod has been accepted by the Kubernetes system, but one or more of the container images has not been created. This includes time before being scheduled as well as time spent downloading images over the network, which could take a while. Running: The pod has been bound to a node, and all of the containers have been created. At least one container is still running, or is in the process of starting or restarting. Succeeded: All containers in the pod have terminated in success, and will not be restarted. Failed: All containers in the pod have terminated, and at least one container has terminated in failure. The container either exited with non-zero status or was terminated by the system. Unknown: For some reason the state of the pod could not be obtained, typically due to an error in communicating with the host of the pod.\n\nMore info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-phase", args=[d.arg(name='phase', type=d.T.string)]), + withPhase(phase): { phase: phase }, + '#withPodIP':: d.fn(help='IP address allocated to the pod. Routable at least within the cluster. Empty if not yet allocated.', args=[d.arg(name='podIP', type=d.T.string)]), + withPodIP(podIP): { podIP: podIP }, + '#withPodIPs':: d.fn(help='podIPs holds the IP addresses allocated to the pod. If this field is specified, the 0th entry must match the podIP field. Pods may be allocated at most 1 value for each of IPv4 and IPv6. This list is empty if no IPs have been allocated yet.', args=[d.arg(name='podIPs', type=d.T.array)]), + withPodIPs(podIPs): { podIPs: if std.isArray(v=podIPs) then podIPs else [podIPs] }, + '#withPodIPsMixin':: d.fn(help='podIPs holds the IP addresses allocated to the pod. If this field is specified, the 0th entry must match the podIP field. Pods may be allocated at most 1 value for each of IPv4 and IPv6. This list is empty if no IPs have been allocated yet.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podIPs', type=d.T.array)]), + withPodIPsMixin(podIPs): { podIPs+: if std.isArray(v=podIPs) then podIPs else [podIPs] }, + '#withQosClass':: d.fn(help='The Quality of Service (QOS) classification assigned to the pod based on resource requirements See PodQOSClass type for available QOS classes More info: https://git.k8s.io/community/contributors/design-proposals/node/resource-qos.md', args=[d.arg(name='qosClass', type=d.T.string)]), + withQosClass(qosClass): { qosClass: qosClass }, + '#withReason':: d.fn(help="A brief CamelCase message indicating details about why the pod is in this state. e.g. 'Evicted'", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withStartTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='startTime', type=d.T.string)]), + withStartTime(startTime): { startTime: startTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplate.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplate.libsonnet new file mode 100644 index 00000000000..bfa3fb4b628 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplate.libsonnet @@ -0,0 +1,295 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podTemplate', url='', help='PodTemplate describes a template for creating copies of a predefined pod.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Podtemplate', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'PodTemplate', + } + self.metadata.withName(name=name), + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplateSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplateSpec.libsonnet new file mode 100644 index 00000000000..c56c6d1a0c1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/podTemplateSpec.libsonnet @@ -0,0 +1,242 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podTemplateSpec', url='', help='PodTemplateSpec describes the data a pod should have when created from a template'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seccompProfile+: { type: type } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { securityContext+: { fsGroup: fsGroup } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { securityContext+: { runAsGroup: runAsGroup } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { securityContext+: { runAsUser: runAsUser } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { automountServiceAccountToken: automountServiceAccountToken } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { dnsPolicy: dnsPolicy } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { enableServiceLinks: enableServiceLinks } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { hostIPC: hostIPC } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { hostNetwork: hostNetwork } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { hostPID: hostPID } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { hostname: hostname } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { nodeName: nodeName } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { nodeSelector+: nodeSelector } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { overhead: overhead } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { overhead+: overhead } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { preemptionPolicy: preemptionPolicy } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { priority: priority } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { restartPolicy: restartPolicy } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { runtimeClassName: runtimeClassName } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { serviceAccount: serviceAccount } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { serviceAccountName: serviceAccountName } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { shareProcessNamespace: shareProcessNamespace } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { subdomain: subdomain } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/portworxVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/portworxVolumeSource.libsonnet new file mode 100644 index 00000000000..223141072ea --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/portworxVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='portworxVolumeSource', url='', help='PortworxVolumeSource represents a Portworx volume resource.'), + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { volumeID: volumeID }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/preferredSchedulingTerm.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/preferredSchedulingTerm.libsonnet new file mode 100644 index 00000000000..ee3d34fdb99 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/preferredSchedulingTerm.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='preferredSchedulingTerm', url='', help="An empty preferred scheduling term matches all objects with implicit weight 0 (i.e. it's a no-op). A null preferred scheduling term matches no objects (i.e. is also a no-op)."), + '#preference':: d.obj(help='A null or empty node selector term matches no objects. The requirements of them are ANDed. The TopologySelectorTerm type implements a subset of the NodeSelectorTerm.'), + preference: { + '#withMatchExpressions':: d.fn(help="A list of node selector requirements by node's labels.", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { preference+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help="A list of node selector requirements by node's labels.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { preference+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchFields':: d.fn(help="A list of node selector requirements by node's fields.", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { preference+: { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + '#withMatchFieldsMixin':: d.fn(help="A list of node selector requirements by node's fields.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { preference+: { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + }, + '#withWeight':: d.fn(help='Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100.', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/probe.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/probe.libsonnet new file mode 100644 index 00000000000..daa7fd972f8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/probe.libsonnet @@ -0,0 +1,45 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='probe', url='', help='Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.'), + '#exec':: d.obj(help='ExecAction describes a "run in container" action.'), + exec: { + '#withCommand':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { exec+: { command: if std.isArray(v=command) then command else [command] } }, + '#withCommandMixin':: d.fn(help="Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { exec+: { command+: if std.isArray(v=command) then command else [command] } }, + }, + '#httpGet':: d.obj(help='HTTPGetAction describes an action based on HTTP Get requests.'), + httpGet: { + '#withHost':: d.fn(help='Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { httpGet+: { host: host } }, + '#withHttpHeaders':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } }, + '#withHttpHeadersMixin':: d.fn(help='Custom headers to set in the request. HTTP allows repeated headers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } }, + '#withPath':: d.fn(help='Path to access on the HTTP server.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { httpGet+: { path: path } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { httpGet+: { port: port } }, + '#withScheme':: d.fn(help='Scheme to use for connecting to the host. Defaults to HTTP.', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { httpGet+: { scheme: scheme } }, + }, + '#tcpSocket':: d.obj(help='TCPSocketAction describes an action based on opening a socket'), + tcpSocket: { + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { tcpSocket+: { host: host } }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { tcpSocket+: { port: port } }, + }, + '#withFailureThreshold':: d.fn(help='Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1.', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { failureThreshold: failureThreshold }, + '#withInitialDelaySeconds':: d.fn(help='Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { initialDelaySeconds: initialDelaySeconds }, + '#withPeriodSeconds':: d.fn(help='How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1.', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { periodSeconds: periodSeconds }, + '#withSuccessThreshold':: d.fn(help='Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1.', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { successThreshold: successThreshold }, + '#withTimeoutSeconds':: d.fn(help='Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { timeoutSeconds: timeoutSeconds }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/projectedVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/projectedVolumeSource.libsonnet new file mode 100644 index 00000000000..e2e272c2db1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/projectedVolumeSource.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='projectedVolumeSource', url='', help='Represents a projected volume source'), + '#withDefaultMode':: d.fn(help='Mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { defaultMode: defaultMode }, + '#withSources':: d.fn(help='list of volume projections', args=[d.arg(name='sources', type=d.T.array)]), + withSources(sources): { sources: if std.isArray(v=sources) then sources else [sources] }, + '#withSourcesMixin':: d.fn(help='list of volume projections\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]), + withSourcesMixin(sources): { sources+: if std.isArray(v=sources) then sources else [sources] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/quobyteVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/quobyteVolumeSource.libsonnet new file mode 100644 index 00000000000..02f20fa3559 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/quobyteVolumeSource.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='quobyteVolumeSource', url='', help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { registry: registry }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { tenant: tenant }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { volume: volume }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..e69a4a147d9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdPersistentVolumeSource.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rbdPersistentVolumeSource', url='', help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { keyring: keyring }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { monitors: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { monitors+: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { pool: pool }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdVolumeSource.libsonnet new file mode 100644 index 00000000000..dea713cf30e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/rbdVolumeSource.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rbdVolumeSource', url='', help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { keyring: keyring }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { monitors: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { monitors+: if std.isArray(v=monitors) then monitors else [monitors] }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { pool: pool }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationController.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationController.libsonnet new file mode 100644 index 00000000000..2a4a2723e18 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationController.libsonnet @@ -0,0 +1,306 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicationController', url='', help='ReplicationController represents the configuration of a replication controller.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Replicationcontroller', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'ReplicationController', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='ReplicationControllerSpec is the specification of a replication controller.'), + spec: { + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { spec+: { template+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { spec+: { template+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { spec+: { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { spec+: { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { spec+: { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { spec+: { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { spec+: { template+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { spec+: { template+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { spec+: { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { spec+: { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { template+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { template+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { spec+: { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { template+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { spec+: { template+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { template+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withReplicas':: d.fn(help='Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + '#withSelector':: d.fn(help='Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors', args=[d.arg(name='selector', type=d.T.object)]), + withSelector(selector): { spec+: { selector: selector } }, + '#withSelectorMixin':: d.fn(help='Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='selector', type=d.T.object)]), + withSelectorMixin(selector): { spec+: { selector+: selector } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerCondition.libsonnet new file mode 100644 index 00000000000..e978d0af96b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicationControllerCondition', url='', help='ReplicationControllerCondition describes the state of a replication controller at a certain point.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='A human readable message indicating details about the transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="The reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='Type of replication controller condition.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerSpec.libsonnet new file mode 100644 index 00000000000..6b283d5a878 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerSpec.libsonnet @@ -0,0 +1,253 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicationControllerSpec', url='', help='ReplicationControllerSpec is the specification of a replication controller.'), + '#template':: d.obj(help='PodTemplateSpec describes the data a pod should have when created from a template'), + template: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { template+: { metadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { template+: { metadata+: { annotations+: annotations } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { template+: { metadata+: { clusterName: clusterName } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { template+: { metadata+: { creationTimestamp: creationTimestamp } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { template+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { template+: { metadata+: { deletionTimestamp: deletionTimestamp } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { template+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { template+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { template+: { metadata+: { generateName: generateName } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { template+: { metadata+: { generation: generation } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { template+: { metadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { template+: { metadata+: { labels+: labels } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { template+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { template+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { template+: { metadata+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { template+: { metadata+: { namespace: namespace } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { template+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { template+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { template+: { metadata+: { resourceVersion: resourceVersion } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { template+: { metadata+: { selfLink: selfLink } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { template+: { metadata+: { uid: uid } } }, + }, + '#spec':: d.obj(help='PodSpec is a description of a pod.'), + spec: { + '#affinity':: d.obj(help='Affinity is a group of affinity scheduling rules.'), + affinity: { + '#nodeAffinity':: d.obj(help='Node affinity is a group of node affinity scheduling rules.'), + nodeAffinity: { + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAffinity':: d.obj(help='Pod affinity is a group of inter pod affinity scheduling rules.'), + podAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + '#podAntiAffinity':: d.obj(help='Pod anti affinity is a group of inter pod anti affinity scheduling rules.'), + podAntiAffinity: { + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } }, + }, + }, + '#dnsConfig':: d.obj(help='PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy.'), + dnsConfig: { + '#withNameservers':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameservers(nameservers): { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withNameserversMixin':: d.fn(help='A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]), + withNameserversMixin(nameservers): { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } }, + '#withOptions':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.', args=[d.arg(name='options', type=d.T.array)]), + withOptions(options): { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } }, + '#withOptionsMixin':: d.fn(help='A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]), + withOptionsMixin(options): { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } }, + '#withSearches':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.', args=[d.arg(name='searches', type=d.T.array)]), + withSearches(searches): { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } }, + '#withSearchesMixin':: d.fn(help='A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]), + withSearchesMixin(searches): { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } }, + }, + '#securityContext':: d.obj(help='PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext.'), + securityContext: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } }, + }, + '#withFsGroup':: d.fn(help="A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod:\n\n1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw----\n\nIf unset, the Kubelet will not modify the ownership and permissions of any volume.", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } }, + '#withFsGroupChangePolicy':: d.fn(help='fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are "OnRootMismatch" and "Always". If not specified defaults to "Always".', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } }, + '#withSupplementalGroups':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSupplementalGroupsMixin':: d.fn(help="A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } }, + '#withSysctls':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + '#withSysctlsMixin':: d.fn(help='Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } }, + }, + '#withActiveDeadlineSeconds':: d.fn(help='Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer.', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]), + withActiveDeadlineSeconds(activeDeadlineSeconds): { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } }, + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether a service account token should be automatically mounted.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } }, + '#withContainers':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } }, + '#withContainersMixin':: d.fn(help='List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } }, + '#withDnsPolicy':: d.fn(help="Set DNS policy for the pod. Defaults to 'ClusterFirst'. Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.", args=[d.arg(name='dnsPolicy', type=d.T.string)]), + withDnsPolicy(dnsPolicy): { template+: { spec+: { dnsPolicy: dnsPolicy } } }, + '#withEnableServiceLinks':: d.fn(help="EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]), + withEnableServiceLinks(enableServiceLinks): { template+: { spec+: { enableServiceLinks: enableServiceLinks } } }, + '#withEphemeralContainers':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainers(ephemeralContainers): { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withEphemeralContainersMixin':: d.fn(help="List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource. This field is alpha-level and is only honored by servers that enable the EphemeralContainers feature.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]), + withEphemeralContainersMixin(ephemeralContainers): { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } }, + '#withHostAliases':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostAliasesMixin':: d.fn(help="HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } }, + '#withHostIPC':: d.fn(help="Use the host's ipc namespace. Optional: Default to false.", args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { template+: { spec+: { hostIPC: hostIPC } } }, + '#withHostNetwork':: d.fn(help="Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { template+: { spec+: { hostNetwork: hostNetwork } } }, + '#withHostPID':: d.fn(help="Use the host's pid namespace. Optional: Default to false.", args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { template+: { spec+: { hostPID: hostPID } } }, + '#withHostname':: d.fn(help="Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.", args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { template+: { spec+: { hostname: hostname } } }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } }, + '#withInitContainers':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withInitContainersMixin':: d.fn(help='List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } }, + '#withNodeName':: d.fn(help='NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { template+: { spec+: { nodeName: nodeName } } }, + '#withNodeSelector':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { template+: { spec+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { template+: { spec+: { nodeSelector+: nodeSelector } } }, + '#withOverhead':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.', args=[d.arg(name='overhead', type=d.T.object)]), + withOverhead(overhead): { template+: { spec+: { overhead: overhead } } }, + '#withOverheadMixin':: d.fn(help='Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/20190226-pod-overhead.md This field is alpha-level as of Kubernetes v1.16, and is only honored by servers that enable the PodOverhead feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]), + withOverheadMixin(overhead): { template+: { spec+: { overhead+: overhead } } }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { template+: { spec+: { preemptionPolicy: preemptionPolicy } } }, + '#withPriority':: d.fn(help='The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority.', args=[d.arg(name='priority', type=d.T.integer)]), + withPriority(priority): { template+: { spec+: { priority: priority } } }, + '#withPriorityClassName':: d.fn(help="If specified, indicates the pod's priority. 'system-node-critical' and 'system-cluster-critical' are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.", args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { template+: { spec+: { priorityClassName: priorityClassName } } }, + '#withReadinessGates':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGates(readinessGates): { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withReadinessGatesMixin':: d.fn(help='If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]), + withReadinessGatesMixin(readinessGates): { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } }, + '#withRestartPolicy':: d.fn(help='Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { template+: { spec+: { restartPolicy: restartPolicy } } }, + '#withRuntimeClassName':: d.fn(help='RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is a beta feature as of Kubernetes v1.14.', args=[d.arg(name='runtimeClassName', type=d.T.string)]), + withRuntimeClassName(runtimeClassName): { template+: { spec+: { runtimeClassName: runtimeClassName } } }, + '#withSchedulerName':: d.fn(help='If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler.', args=[d.arg(name='schedulerName', type=d.T.string)]), + withSchedulerName(schedulerName): { template+: { spec+: { schedulerName: schedulerName } } }, + '#withServiceAccount':: d.fn(help='DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead.', args=[d.arg(name='serviceAccount', type=d.T.string)]), + withServiceAccount(serviceAccount): { template+: { spec+: { serviceAccount: serviceAccount } } }, + '#withServiceAccountName':: d.fn(help='ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { template+: { spec+: { serviceAccountName: serviceAccountName } } }, + '#withSetHostnameAsFQDN':: d.fn(help="If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\SYSTEM\\CurrentControlSet\\Services\\Tcpip\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]), + withSetHostnameAsFQDN(setHostnameAsFQDN): { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } }, + '#withShareProcessNamespace':: d.fn(help='Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false.', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]), + withShareProcessNamespace(shareProcessNamespace): { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } }, + '#withSubdomain':: d.fn(help='If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all.', args=[d.arg(name='subdomain', type=d.T.string)]), + withSubdomain(subdomain): { template+: { spec+: { subdomain: subdomain } } }, + '#withTerminationGracePeriodSeconds':: d.fn(help='Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds.', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } }, + '#withTolerations':: d.fn(help="If specified, the pod's tolerations.", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help="If specified, the pod's tolerations.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTopologySpreadConstraints':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help='TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } }, + '#withVolumes':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } }, + '#withVolumesMixin':: d.fn(help='List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } }, + }, + }, + '#withMinReadySeconds':: d.fn(help='Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready)', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { minReadySeconds: minReadySeconds }, + '#withReplicas':: d.fn(help='Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#withSelector':: d.fn(help='Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors', args=[d.arg(name='selector', type=d.T.object)]), + withSelector(selector): { selector: selector }, + '#withSelectorMixin':: d.fn(help='Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='selector', type=d.T.object)]), + withSelectorMixin(selector): { selector+: selector }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerStatus.libsonnet new file mode 100644 index 00000000000..8e0121b7a91 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/replicationControllerStatus.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='replicationControllerStatus', url='', help='ReplicationControllerStatus represents the current status of a replication controller.'), + '#withAvailableReplicas':: d.fn(help='The number of available replicas (ready for at least minReadySeconds) for this replication controller.', args=[d.arg(name='availableReplicas', type=d.T.integer)]), + withAvailableReplicas(availableReplicas): { availableReplicas: availableReplicas }, + '#withConditions':: d.fn(help="Represents the latest available observations of a replication controller's current state.", args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help="Represents the latest available observations of a replication controller's current state.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withFullyLabeledReplicas':: d.fn(help='The number of pods that have labels matching the labels of the pod template of the replication controller.', args=[d.arg(name='fullyLabeledReplicas', type=d.T.integer)]), + withFullyLabeledReplicas(fullyLabeledReplicas): { fullyLabeledReplicas: fullyLabeledReplicas }, + '#withObservedGeneration':: d.fn(help='ObservedGeneration reflects the generation of the most recently observed replication controller.', args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#withReadyReplicas':: d.fn(help='The number of ready replicas for this replication controller.', args=[d.arg(name='readyReplicas', type=d.T.integer)]), + withReadyReplicas(readyReplicas): { readyReplicas: readyReplicas }, + '#withReplicas':: d.fn(help='Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { replicas: replicas }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceFieldSelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceFieldSelector.libsonnet new file mode 100644 index 00000000000..506aadf4533 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceFieldSelector.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceFieldSelector', url='', help='ResourceFieldSelector represents container resources (cpu, memory) and their output format'), + '#withContainerName':: d.fn(help='Container name: required for volumes, optional for env vars', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { containerName: containerName }, + '#withDivisor':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='divisor', type=d.T.string)]), + withDivisor(divisor): { divisor: divisor }, + '#withResource':: d.fn(help='Required: resource to select', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resource: resource }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuota.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuota.libsonnet new file mode 100644 index 00000000000..4bbc27ef31d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuota.libsonnet @@ -0,0 +1,74 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceQuota', url='', help='ResourceQuota sets aggregate quota restrictions enforced per namespace'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Resourcequota', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'ResourceQuota', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='ResourceQuotaSpec defines the desired hard limits to enforce for Quota.'), + spec: { + '#scopeSelector':: d.obj(help='A scope selector represents the AND of the selectors represented by the scoped-resource selector requirements.'), + scopeSelector: { + '#withMatchExpressions':: d.fn(help='A list of scope selector requirements by scope of the resources.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { scopeSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='A list of scope selector requirements by scope of the resources.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { scopeSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + }, + '#withHard':: d.fn(help='hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/', args=[d.arg(name='hard', type=d.T.object)]), + withHard(hard): { spec+: { hard: hard } }, + '#withHardMixin':: d.fn(help='hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hard', type=d.T.object)]), + withHardMixin(hard): { spec+: { hard+: hard } }, + '#withScopes':: d.fn(help='A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects.', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { spec+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withScopesMixin':: d.fn(help='A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { spec+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaSpec.libsonnet new file mode 100644 index 00000000000..8af3a38d93d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaSpec.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceQuotaSpec', url='', help='ResourceQuotaSpec defines the desired hard limits to enforce for Quota.'), + '#scopeSelector':: d.obj(help='A scope selector represents the AND of the selectors represented by the scoped-resource selector requirements.'), + scopeSelector: { + '#withMatchExpressions':: d.fn(help='A list of scope selector requirements by scope of the resources.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { scopeSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='A list of scope selector requirements by scope of the resources.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { scopeSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + }, + '#withHard':: d.fn(help='hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/', args=[d.arg(name='hard', type=d.T.object)]), + withHard(hard): { hard: hard }, + '#withHardMixin':: d.fn(help='hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hard', type=d.T.object)]), + withHardMixin(hard): { hard+: hard }, + '#withScopes':: d.fn(help='A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects.', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { scopes: if std.isArray(v=scopes) then scopes else [scopes] }, + '#withScopesMixin':: d.fn(help='A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { scopes+: if std.isArray(v=scopes) then scopes else [scopes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaStatus.libsonnet new file mode 100644 index 00000000000..f61db494f1b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceQuotaStatus.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceQuotaStatus', url='', help='ResourceQuotaStatus defines the enforced hard limits and observed use.'), + '#withHard':: d.fn(help='Hard is the set of enforced hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/', args=[d.arg(name='hard', type=d.T.object)]), + withHard(hard): { hard: hard }, + '#withHardMixin':: d.fn(help='Hard is the set of enforced hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hard', type=d.T.object)]), + withHardMixin(hard): { hard+: hard }, + '#withUsed':: d.fn(help='Used is the current observed total usage of the resource in the namespace.', args=[d.arg(name='used', type=d.T.object)]), + withUsed(used): { used: used }, + '#withUsedMixin':: d.fn(help='Used is the current observed total usage of the resource in the namespace.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='used', type=d.T.object)]), + withUsedMixin(used): { used+: used }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceRequirements.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceRequirements.libsonnet new file mode 100644 index 00000000000..31db43ac28d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/resourceRequirements.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourceRequirements', url='', help='ResourceRequirements describes the compute resource requirements.'), + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { limits: limits }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { limits+: limits }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { requests: requests }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { requests+: requests }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..0526d2a1df5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOPersistentVolumeSource.libsonnet @@ -0,0 +1,31 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scaleIOPersistentVolumeSource', url='', help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { gateway: gateway }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { protectionDomain: protectionDomain }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { sslEnabled: sslEnabled }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { storageMode: storageMode }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { storagePool: storagePool }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { system: system }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeName: volumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOVolumeSource.libsonnet new file mode 100644 index 00000000000..b596bf11bbc --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scaleIOVolumeSource.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scaleIOVolumeSource', url='', help='ScaleIOVolumeSource represents a persistent ScaleIO volume'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { gateway: gateway }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { protectionDomain: protectionDomain }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { sslEnabled: sslEnabled }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { storageMode: storageMode }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { storagePool: storagePool }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { system: system }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeName: volumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopeSelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopeSelector.libsonnet new file mode 100644 index 00000000000..e48c3bc11a7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopeSelector.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scopeSelector', url='', help='A scope selector represents the AND of the selectors represented by the scoped-resource selector requirements.'), + '#withMatchExpressions':: d.fn(help='A list of scope selector requirements by scope of the resources.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchExpressionsMixin':: d.fn(help='A list of scope selector requirements by scope of the resources.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopedResourceSelectorRequirement.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopedResourceSelectorRequirement.libsonnet new file mode 100644 index 00000000000..2c5854532a8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/scopedResourceSelectorRequirement.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scopedResourceSelectorRequirement', url='', help='A scoped-resource selector requirement is a selector that contains values, a scope name, and an operator that relates the scope name and values.'), + '#withOperator':: d.fn(help="Represents a scope's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist.", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withScopeName':: d.fn(help='The name of the scope that the selector applies to.', args=[d.arg(name='scopeName', type=d.T.string)]), + withScopeName(scopeName): { scopeName: scopeName }, + '#withValues':: d.fn(help='An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seLinuxOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seLinuxOptions.libsonnet new file mode 100644 index 00000000000..8b52aac9625 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seLinuxOptions.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='seLinuxOptions', url='', help='SELinuxOptions are the labels to be applied to the container'), + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { level: level }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { role: role }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { user: user }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seccompProfile.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seccompProfile.libsonnet new file mode 100644 index 00000000000..13ee88dd9d3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/seccompProfile.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='seccompProfile', url='', help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { localhostProfile: localhostProfile }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secret.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secret.libsonnet new file mode 100644 index 00000000000..4d9ca0f7dbe --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secret.libsonnet @@ -0,0 +1,68 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secret', url='', help='Secret holds secret data of a certain type. The total bytes of the values in the Data field must be less than MaxSecretSize bytes.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Secret', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Secret', + } + self.metadata.withName(name=name), + '#withData':: d.fn(help="Data contains the secret data. Each key must consist of alphanumeric characters, '-', '_' or '.'. The serialized form of the secret data is a base64 encoded string, representing the arbitrary (possibly non-string) data value here. Described in https://tools.ietf.org/html/rfc4648#section-4", args=[d.arg(name='data', type=d.T.object)]), + withData(data): { data: data }, + '#withDataMixin':: d.fn(help="Data contains the secret data. Each key must consist of alphanumeric characters, '-', '_' or '.'. The serialized form of the secret data is a base64 encoded string, representing the arbitrary (possibly non-string) data value here. Described in https://tools.ietf.org/html/rfc4648#section-4\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='data', type=d.T.object)]), + withDataMixin(data): { data+: data }, + '#withImmutable':: d.fn(help='Immutable, if set to true, ensures that data stored in the Secret cannot be updated (only object metadata can be modified). If not set to true, the field can be modified at any time. Defaulted to nil. This is a beta field enabled by ImmutableEphemeralVolumes feature gate.', args=[d.arg(name='immutable', type=d.T.boolean)]), + withImmutable(immutable): { immutable: immutable }, + '#withStringData':: d.fn(help='stringData allows specifying non-binary secret data in string form. It is provided as a write-only convenience method. All keys and values are merged into the data field on write, overwriting any existing values. It is never output when reading from the API.', args=[d.arg(name='stringData', type=d.T.object)]), + withStringData(stringData): { stringData: stringData }, + '#withStringDataMixin':: d.fn(help='stringData allows specifying non-binary secret data in string form. It is provided as a write-only convenience method. All keys and values are merged into the data field on write, overwriting any existing values. It is never output when reading from the API.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='stringData', type=d.T.object)]), + withStringDataMixin(stringData): { stringData+: stringData }, + '#withType':: d.fn(help='Used to facilitate programmatic handling of secret data.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretEnvSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretEnvSource.libsonnet new file mode 100644 index 00000000000..cb7600a39a6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretEnvSource.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secretEnvSource', url='', help="SecretEnvSource selects a Secret to populate the environment variables with.\n\nThe contents of the target Secret's Data field will represent the key-value pairs as environment variables."), + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the Secret must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretKeySelector.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretKeySelector.libsonnet new file mode 100644 index 00000000000..2a2d79e9552 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretKeySelector.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secretKeySelector', url='', help='SecretKeySelector selects a key of a Secret.'), + '#withKey':: d.fn(help='The key of the secret to select from. Must be a valid secret key.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the Secret or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretProjection.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretProjection.libsonnet new file mode 100644 index 00000000000..2b615bf87d0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretProjection.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secretProjection', url='', help="Adapts a secret into a projected volume.\n\nThe contents of the target Secret's Data field will be presented in a projected volume as files using the keys in the Data field as the file names. Note that this is identical to a secret volume source without the default mode."), + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withOptional':: d.fn(help='Specify whether the Secret or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretReference.libsonnet new file mode 100644 index 00000000000..c0bda20e032 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretReference.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secretReference', url='', help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretVolumeSource.libsonnet new file mode 100644 index 00000000000..5229fd3ba94 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/secretVolumeSource.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='secretVolumeSource', url='', help="Adapts a Secret into a volume.\n\nThe contents of the target Secret's Data field will be presented in a volume as files using the keys in the Data field as the file names. Secret volumes support ownership management and SELinux relabeling."), + '#withDefaultMode':: d.fn(help='Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { defaultMode: defaultMode }, + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { items: if std.isArray(v=items) then items else [items] }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { items+: if std.isArray(v=items) then items else [items] }, + '#withOptional':: d.fn(help='Specify whether the Secret or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { optional: optional }, + '#withSecretName':: d.fn(help="Name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret", args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/securityContext.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/securityContext.libsonnet new file mode 100644 index 00000000000..5b1bb37c28f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/securityContext.libsonnet @@ -0,0 +1,58 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='securityContext', url='', help='SecurityContext holds security configuration that will be applied to a container. Some fields are present in both SecurityContext and PodSecurityContext. When both are set, the values in SecurityContext take precedence.'), + '#capabilities':: d.obj(help='Adds and removes POSIX capabilities from running containers.'), + capabilities: { + '#withAdd':: d.fn(help='Added capabilities', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { capabilities+: { add: if std.isArray(v=add) then add else [add] } }, + '#withAddMixin':: d.fn(help='Added capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { capabilities+: { add+: if std.isArray(v=add) then add else [add] } }, + '#withDrop':: d.fn(help='Removed capabilities', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } }, + '#withDropMixin':: d.fn(help='Removed capabilities\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } }, + }, + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { seLinuxOptions+: { level: level } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { seLinuxOptions+: { role: role } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seLinuxOptions+: { type: type } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { seLinuxOptions+: { user: user } }, + }, + '#seccompProfile':: d.obj(help="SeccompProfile defines a pod/container's seccomp profile settings. Only one profile source may be set."), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is 'Localhost'.", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { seccompProfile+: { localhostProfile: localhostProfile } }, + '#withType':: d.fn(help='type indicates which kind of seccomp profile will be applied. Valid options are:\n\nLocalhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seccompProfile+: { type: type } }, + }, + '#windowsOptions':: d.obj(help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { windowsOptions+: { runAsUserName: runAsUserName } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { allowPrivilegeEscalation: allowPrivilegeEscalation }, + '#withPrivileged':: d.fn(help='Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false.', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { privileged: privileged }, + '#withProcMount':: d.fn(help='procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled.', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { procMount: procMount }, + '#withReadOnlyRootFilesystem':: d.fn(help='Whether this container has a read-only root filesystem. Default is false.', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { readOnlyRootFilesystem: readOnlyRootFilesystem }, + '#withRunAsGroup':: d.fn(help='The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { runAsGroup: runAsGroup }, + '#withRunAsNonRoot':: d.fn(help='Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { runAsNonRoot: runAsNonRoot }, + '#withRunAsUser':: d.fn(help='The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { runAsUser: runAsUser }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/service.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/service.libsonnet new file mode 100644 index 00000000000..adb922e0519 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/service.libsonnet @@ -0,0 +1,105 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='service', url='', help='Service is a named abstraction of software service (for example, mysql) consisting of local port (for example 3306) that the proxy listens on, and the selector that determines which pods will answer requests sent through the proxy.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Service', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'Service', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='ServiceSpec describes the attributes that a user creates on a service.'), + spec: { + '#sessionAffinityConfig':: d.obj(help='SessionAffinityConfig represents the configurations of session affinity.'), + sessionAffinityConfig: { + '#clientIP':: d.obj(help='ClientIPConfig represents the configurations of Client IP based session affinity.'), + clientIP: { + '#withTimeoutSeconds':: d.fn(help='timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours).', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { spec+: { sessionAffinityConfig+: { clientIP+: { timeoutSeconds: timeoutSeconds } } } }, + }, + }, + '#withClusterIP':: d.fn(help='clusterIP is the IP address of the service and is usually assigned randomly by the master. If an address is specified manually and is not in use by others, it will be allocated to the service; otherwise, creation of the service will fail. This field can not be changed through updates. Valid values are "None", empty string (""), or a valid IP address. "None" can be specified for headless services when proxying is not required. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='clusterIP', type=d.T.string)]), + withClusterIP(clusterIP): { spec+: { clusterIP: clusterIP } }, + '#withExternalIPs':: d.fn(help='externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system.', args=[d.arg(name='externalIPs', type=d.T.array)]), + withExternalIPs(externalIPs): { spec+: { externalIPs: if std.isArray(v=externalIPs) then externalIPs else [externalIPs] } }, + '#withExternalIPsMixin':: d.fn(help='externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='externalIPs', type=d.T.array)]), + withExternalIPsMixin(externalIPs): { spec+: { externalIPs+: if std.isArray(v=externalIPs) then externalIPs else [externalIPs] } }, + '#withExternalName':: d.fn(help='externalName is the external reference that kubedns or equivalent will return as a CNAME record for this service. No proxying will be involved. Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123) and requires Type to be ExternalName.', args=[d.arg(name='externalName', type=d.T.string)]), + withExternalName(externalName): { spec+: { externalName: externalName } }, + '#withExternalTrafficPolicy':: d.fn(help='externalTrafficPolicy denotes if this Service desires to route external traffic to node-local or cluster-wide endpoints. "Local" preserves the client source IP and avoids a second hop for LoadBalancer and Nodeport type services, but risks potentially imbalanced traffic spreading. "Cluster" obscures the client source IP and may cause a second hop to another node, but should have good overall load-spreading.', args=[d.arg(name='externalTrafficPolicy', type=d.T.string)]), + withExternalTrafficPolicy(externalTrafficPolicy): { spec+: { externalTrafficPolicy: externalTrafficPolicy } }, + '#withHealthCheckNodePort':: d.fn(help='healthCheckNodePort specifies the healthcheck nodePort for the service. If not specified, HealthCheckNodePort is created by the service api backend with the allocated nodePort. Will use user-specified nodePort value if specified by the client. Only effects when Type is set to LoadBalancer and ExternalTrafficPolicy is set to Local.', args=[d.arg(name='healthCheckNodePort', type=d.T.integer)]), + withHealthCheckNodePort(healthCheckNodePort): { spec+: { healthCheckNodePort: healthCheckNodePort } }, + '#withIpFamily':: d.fn(help="ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. IPv6) when the IPv6DualStack feature gate is enabled. In a dual-stack cluster, you can specify ipFamily when creating a ClusterIP Service to determine whether the controller will allocate an IPv4 or IPv6 IP for it, and you can specify ipFamily when creating a headless Service to determine whether it will have IPv4 or IPv6 Endpoints. In either case, if you do not specify an ipFamily explicitly, it will default to the cluster's primary IP family. This field is part of an alpha feature, and you should not make any assumptions about its semantics other than those described above. In particular, you should not assume that it can (or cannot) be changed after creation time; that it can only have the values 'IPv4' and 'IPv6'; or that its current value on a given Service correctly reflects the current state of that Service. (For ClusterIP Services, look at clusterIP to see if the Service is IPv4 or IPv6. For headless Services, look at the endpoints, which may be dual-stack in the future. For ExternalName Services, ipFamily has no meaning, but it may be set to an irrelevant value anyway.)", args=[d.arg(name='ipFamily', type=d.T.string)]), + withIpFamily(ipFamily): { spec+: { ipFamily: ipFamily } }, + '#withLoadBalancerIP':: d.fn(help='Only applies to Service Type: LoadBalancer LoadBalancer will get created with the IP specified in this field. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature.', args=[d.arg(name='loadBalancerIP', type=d.T.string)]), + withLoadBalancerIP(loadBalancerIP): { spec+: { loadBalancerIP: loadBalancerIP } }, + '#withLoadBalancerSourceRanges':: d.fn(help='If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/', args=[d.arg(name='loadBalancerSourceRanges', type=d.T.array)]), + withLoadBalancerSourceRanges(loadBalancerSourceRanges): { spec+: { loadBalancerSourceRanges: if std.isArray(v=loadBalancerSourceRanges) then loadBalancerSourceRanges else [loadBalancerSourceRanges] } }, + '#withLoadBalancerSourceRangesMixin':: d.fn(help='If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='loadBalancerSourceRanges', type=d.T.array)]), + withLoadBalancerSourceRangesMixin(loadBalancerSourceRanges): { spec+: { loadBalancerSourceRanges+: if std.isArray(v=loadBalancerSourceRanges) then loadBalancerSourceRanges else [loadBalancerSourceRanges] } }, + '#withPorts':: d.fn(help='The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { spec+: { ports: if std.isArray(v=ports) then ports else [ports] } }, + '#withPortsMixin':: d.fn(help='The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { spec+: { ports+: if std.isArray(v=ports) then ports else [ports] } }, + '#withPublishNotReadyAddresses':: d.fn(help="publishNotReadyAddresses indicates that any agent which deals with endpoints for this Service should disregard any indications of ready/not-ready. The primary use case for setting this field is for a StatefulSet's Headless Service to propagate SRV DNS records for its Pods for the purpose of peer discovery. The Kubernetes controllers that generate Endpoints and EndpointSlice resources for Services interpret this to mean that all endpoints are considered 'ready' even if the Pods themselves are not. Agents which consume only Kubernetes generated endpoints through the Endpoints or EndpointSlice resources can safely assume this behavior.", args=[d.arg(name='publishNotReadyAddresses', type=d.T.boolean)]), + withPublishNotReadyAddresses(publishNotReadyAddresses): { spec+: { publishNotReadyAddresses: publishNotReadyAddresses } }, + '#withSelector':: d.fn(help='Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/', args=[d.arg(name='selector', type=d.T.object)]), + withSelector(selector): { spec+: { selector: selector } }, + '#withSelectorMixin':: d.fn(help='Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='selector', type=d.T.object)]), + withSelectorMixin(selector): { spec+: { selector+: selector } }, + '#withSessionAffinity':: d.fn(help='Supports "ClientIP" and "None". Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='sessionAffinity', type=d.T.string)]), + withSessionAffinity(sessionAffinity): { spec+: { sessionAffinity: sessionAffinity } }, + '#withTopologyKeys':: d.fn(help='topologyKeys is a preference-order list of topology keys which implementations of services should use to preferentially sort endpoints when accessing this Service, it can not be used at the same time as externalTrafficPolicy=Local. Topology keys must be valid label keys and at most 16 keys may be specified. Endpoints are chosen based on the first topology key with available backends. If this field is specified and all entries have no backends that match the topology of the client, the service has no backends for that client and connections should fail. The special value "*" may be used to mean "any topology". This catch-all value, if used, only makes sense as the last value in the list. If this is not specified or empty, no topology constraints will be applied.', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeys(topologyKeys): { spec+: { topologyKeys: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] } }, + '#withTopologyKeysMixin':: d.fn(help='topologyKeys is a preference-order list of topology keys which implementations of services should use to preferentially sort endpoints when accessing this Service, it can not be used at the same time as externalTrafficPolicy=Local. Topology keys must be valid label keys and at most 16 keys may be specified. Endpoints are chosen based on the first topology key with available backends. If this field is specified and all entries have no backends that match the topology of the client, the service has no backends for that client and connections should fail. The special value "*" may be used to mean "any topology". This catch-all value, if used, only makes sense as the last value in the list. If this is not specified or empty, no topology constraints will be applied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeysMixin(topologyKeys): { spec+: { topologyKeys+: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] } }, + '#withType':: d.fn(help='type determines how the Service is exposed. Defaults to ClusterIP. Valid options are ExternalName, ClusterIP, NodePort, and LoadBalancer. "ExternalName" maps to the specified externalName. "ClusterIP" allocates a cluster-internal IP address for load-balancing to endpoints. Endpoints are determined by the selector or if that is not specified, by manual construction of an Endpoints object. If clusterIP is "None", no virtual IP is allocated and the endpoints are published as a set of endpoints rather than a stable IP. "NodePort" builds on ClusterIP and allocates a port on every node which routes to the clusterIP. "LoadBalancer" builds on NodePort and creates an external load-balancer (if supported in the current cloud) which routes to the clusterIP. More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { type: type } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccount.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccount.libsonnet new file mode 100644 index 00000000000..f6b47c99fd2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccount.libsonnet @@ -0,0 +1,66 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceAccount', url='', help='ServiceAccount binds together: * a name, understood by users, and perhaps by peripheral systems, for an identity * a principal that can be authenticated and authorized * a set of secrets'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Serviceaccount', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'v1', + kind: 'ServiceAccount', + } + self.metadata.withName(name=name), + '#withAutomountServiceAccountToken':: d.fn(help='AutomountServiceAccountToken indicates whether pods running as this service account should have an API token automatically mounted. Can be overridden at the pod level.', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { automountServiceAccountToken: automountServiceAccountToken }, + '#withImagePullSecrets':: d.fn(help='ImagePullSecrets is a list of references to secrets in the same namespace to use for pulling any images in pods that reference this ServiceAccount. ImagePullSecrets are distinct from Secrets because Secrets can be mounted in the pod, but ImagePullSecrets are only accessed by the kubelet. More info: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] }, + '#withImagePullSecretsMixin':: d.fn(help='ImagePullSecrets is a list of references to secrets in the same namespace to use for pulling any images in pods that reference this ServiceAccount. ImagePullSecrets are distinct from Secrets because Secrets can be mounted in the pod, but ImagePullSecrets are only accessed by the kubelet. More info: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] }, + '#withSecrets':: d.fn(help='Secrets is the list of secrets allowed to be used by pods running using this ServiceAccount. More info: https://kubernetes.io/docs/concepts/configuration/secret', args=[d.arg(name='secrets', type=d.T.array)]), + withSecrets(secrets): { secrets: if std.isArray(v=secrets) then secrets else [secrets] }, + '#withSecretsMixin':: d.fn(help='Secrets is the list of secrets allowed to be used by pods running using this ServiceAccount. More info: https://kubernetes.io/docs/concepts/configuration/secret\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='secrets', type=d.T.array)]), + withSecretsMixin(secrets): { secrets+: if std.isArray(v=secrets) then secrets else [secrets] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccountTokenProjection.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccountTokenProjection.libsonnet new file mode 100644 index 00000000000..52b7c89e793 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceAccountTokenProjection.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceAccountTokenProjection', url='', help='ServiceAccountTokenProjection represents a projected service account token volume. This projection can be used to insert a service account token into the pods runtime filesystem for use against APIs (Kubernetes API Server or otherwise).'), + '#withAudience':: d.fn(help='Audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver.', args=[d.arg(name='audience', type=d.T.string)]), + withAudience(audience): { audience: audience }, + '#withExpirationSeconds':: d.fn(help='ExpirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes.', args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { expirationSeconds: expirationSeconds }, + '#withPath':: d.fn(help='Path is the path relative to the mount point of the file to project the token into.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/servicePort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/servicePort.libsonnet new file mode 100644 index 00000000000..ddbf94d2583 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/servicePort.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='servicePort', url='', help="ServicePort contains information on service's port."), + '#withAppProtocol':: d.fn(help='The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol. This is a beta field that is guarded by the ServiceAppProtocol feature gate and enabled by default.', args=[d.arg(name='appProtocol', type=d.T.string)]), + withAppProtocol(appProtocol): { appProtocol: appProtocol }, + '#withName':: d.fn(help="The name of this port within the service. This must be a DNS_LABEL. All ports within a ServiceSpec must have unique names. When considering the endpoints for a Service, this must match the 'name' field in the EndpointPort. Optional if only one ServicePort is defined on this service.", args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNodePort':: d.fn(help='The port on each node on which this service is exposed when type=NodePort or LoadBalancer. Usually assigned by the system. If specified, it will be allocated to the service if unused or else creation of the service will fail. Default is to auto-allocate a port if the ServiceType of this Service requires one. More info: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport', args=[d.arg(name='nodePort', type=d.T.integer)]), + withNodePort(nodePort): { nodePort: nodePort }, + '#withPort':: d.fn(help='The port that will be exposed by this service.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#withProtocol':: d.fn(help='The IP protocol for this port. Supports "TCP", "UDP", and "SCTP". Default is TCP.', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + '#withTargetPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='targetPort', type=d.T.string)]), + withTargetPort(targetPort): { targetPort: targetPort }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceSpec.libsonnet new file mode 100644 index 00000000000..0ff53a68769 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceSpec.libsonnet @@ -0,0 +1,52 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceSpec', url='', help='ServiceSpec describes the attributes that a user creates on a service.'), + '#sessionAffinityConfig':: d.obj(help='SessionAffinityConfig represents the configurations of session affinity.'), + sessionAffinityConfig: { + '#clientIP':: d.obj(help='ClientIPConfig represents the configurations of Client IP based session affinity.'), + clientIP: { + '#withTimeoutSeconds':: d.fn(help='timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours).', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { sessionAffinityConfig+: { clientIP+: { timeoutSeconds: timeoutSeconds } } }, + }, + }, + '#withClusterIP':: d.fn(help='clusterIP is the IP address of the service and is usually assigned randomly by the master. If an address is specified manually and is not in use by others, it will be allocated to the service; otherwise, creation of the service will fail. This field can not be changed through updates. Valid values are "None", empty string (""), or a valid IP address. "None" can be specified for headless services when proxying is not required. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='clusterIP', type=d.T.string)]), + withClusterIP(clusterIP): { clusterIP: clusterIP }, + '#withExternalIPs':: d.fn(help='externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system.', args=[d.arg(name='externalIPs', type=d.T.array)]), + withExternalIPs(externalIPs): { externalIPs: if std.isArray(v=externalIPs) then externalIPs else [externalIPs] }, + '#withExternalIPsMixin':: d.fn(help='externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='externalIPs', type=d.T.array)]), + withExternalIPsMixin(externalIPs): { externalIPs+: if std.isArray(v=externalIPs) then externalIPs else [externalIPs] }, + '#withExternalName':: d.fn(help='externalName is the external reference that kubedns or equivalent will return as a CNAME record for this service. No proxying will be involved. Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123) and requires Type to be ExternalName.', args=[d.arg(name='externalName', type=d.T.string)]), + withExternalName(externalName): { externalName: externalName }, + '#withExternalTrafficPolicy':: d.fn(help='externalTrafficPolicy denotes if this Service desires to route external traffic to node-local or cluster-wide endpoints. "Local" preserves the client source IP and avoids a second hop for LoadBalancer and Nodeport type services, but risks potentially imbalanced traffic spreading. "Cluster" obscures the client source IP and may cause a second hop to another node, but should have good overall load-spreading.', args=[d.arg(name='externalTrafficPolicy', type=d.T.string)]), + withExternalTrafficPolicy(externalTrafficPolicy): { externalTrafficPolicy: externalTrafficPolicy }, + '#withHealthCheckNodePort':: d.fn(help='healthCheckNodePort specifies the healthcheck nodePort for the service. If not specified, HealthCheckNodePort is created by the service api backend with the allocated nodePort. Will use user-specified nodePort value if specified by the client. Only effects when Type is set to LoadBalancer and ExternalTrafficPolicy is set to Local.', args=[d.arg(name='healthCheckNodePort', type=d.T.integer)]), + withHealthCheckNodePort(healthCheckNodePort): { healthCheckNodePort: healthCheckNodePort }, + '#withIpFamily':: d.fn(help="ipFamily specifies whether this Service has a preference for a particular IP family (e.g. IPv4 vs. IPv6) when the IPv6DualStack feature gate is enabled. In a dual-stack cluster, you can specify ipFamily when creating a ClusterIP Service to determine whether the controller will allocate an IPv4 or IPv6 IP for it, and you can specify ipFamily when creating a headless Service to determine whether it will have IPv4 or IPv6 Endpoints. In either case, if you do not specify an ipFamily explicitly, it will default to the cluster's primary IP family. This field is part of an alpha feature, and you should not make any assumptions about its semantics other than those described above. In particular, you should not assume that it can (or cannot) be changed after creation time; that it can only have the values 'IPv4' and 'IPv6'; or that its current value on a given Service correctly reflects the current state of that Service. (For ClusterIP Services, look at clusterIP to see if the Service is IPv4 or IPv6. For headless Services, look at the endpoints, which may be dual-stack in the future. For ExternalName Services, ipFamily has no meaning, but it may be set to an irrelevant value anyway.)", args=[d.arg(name='ipFamily', type=d.T.string)]), + withIpFamily(ipFamily): { ipFamily: ipFamily }, + '#withLoadBalancerIP':: d.fn(help='Only applies to Service Type: LoadBalancer LoadBalancer will get created with the IP specified in this field. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature.', args=[d.arg(name='loadBalancerIP', type=d.T.string)]), + withLoadBalancerIP(loadBalancerIP): { loadBalancerIP: loadBalancerIP }, + '#withLoadBalancerSourceRanges':: d.fn(help='If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/', args=[d.arg(name='loadBalancerSourceRanges', type=d.T.array)]), + withLoadBalancerSourceRanges(loadBalancerSourceRanges): { loadBalancerSourceRanges: if std.isArray(v=loadBalancerSourceRanges) then loadBalancerSourceRanges else [loadBalancerSourceRanges] }, + '#withLoadBalancerSourceRangesMixin':: d.fn(help='If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='loadBalancerSourceRanges', type=d.T.array)]), + withLoadBalancerSourceRangesMixin(loadBalancerSourceRanges): { loadBalancerSourceRanges+: if std.isArray(v=loadBalancerSourceRanges) then loadBalancerSourceRanges else [loadBalancerSourceRanges] }, + '#withPorts':: d.fn(help='The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withPublishNotReadyAddresses':: d.fn(help="publishNotReadyAddresses indicates that any agent which deals with endpoints for this Service should disregard any indications of ready/not-ready. The primary use case for setting this field is for a StatefulSet's Headless Service to propagate SRV DNS records for its Pods for the purpose of peer discovery. The Kubernetes controllers that generate Endpoints and EndpointSlice resources for Services interpret this to mean that all endpoints are considered 'ready' even if the Pods themselves are not. Agents which consume only Kubernetes generated endpoints through the Endpoints or EndpointSlice resources can safely assume this behavior.", args=[d.arg(name='publishNotReadyAddresses', type=d.T.boolean)]), + withPublishNotReadyAddresses(publishNotReadyAddresses): { publishNotReadyAddresses: publishNotReadyAddresses }, + '#withSelector':: d.fn(help='Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/', args=[d.arg(name='selector', type=d.T.object)]), + withSelector(selector): { selector: selector }, + '#withSelectorMixin':: d.fn(help='Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='selector', type=d.T.object)]), + withSelectorMixin(selector): { selector+: selector }, + '#withSessionAffinity':: d.fn(help='Supports "ClientIP" and "None". Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies', args=[d.arg(name='sessionAffinity', type=d.T.string)]), + withSessionAffinity(sessionAffinity): { sessionAffinity: sessionAffinity }, + '#withTopologyKeys':: d.fn(help='topologyKeys is a preference-order list of topology keys which implementations of services should use to preferentially sort endpoints when accessing this Service, it can not be used at the same time as externalTrafficPolicy=Local. Topology keys must be valid label keys and at most 16 keys may be specified. Endpoints are chosen based on the first topology key with available backends. If this field is specified and all entries have no backends that match the topology of the client, the service has no backends for that client and connections should fail. The special value "*" may be used to mean "any topology". This catch-all value, if used, only makes sense as the last value in the list. If this is not specified or empty, no topology constraints will be applied.', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeys(topologyKeys): { topologyKeys: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#withTopologyKeysMixin':: d.fn(help='topologyKeys is a preference-order list of topology keys which implementations of services should use to preferentially sort endpoints when accessing this Service, it can not be used at the same time as externalTrafficPolicy=Local. Topology keys must be valid label keys and at most 16 keys may be specified. Endpoints are chosen based on the first topology key with available backends. If this field is specified and all entries have no backends that match the topology of the client, the service has no backends for that client and connections should fail. The special value "*" may be used to mean "any topology". This catch-all value, if used, only makes sense as the last value in the list. If this is not specified or empty, no topology constraints will be applied.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeysMixin(topologyKeys): { topologyKeys+: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#withType':: d.fn(help='type determines how the Service is exposed. Defaults to ClusterIP. Valid options are ExternalName, ClusterIP, NodePort, and LoadBalancer. "ExternalName" maps to the specified externalName. "ClusterIP" allocates a cluster-internal IP address for load-balancing to endpoints. Endpoints are determined by the selector or if that is not specified, by manual construction of an Endpoints object. If clusterIP is "None", no virtual IP is allocated and the endpoints are published as a set of endpoints rather than a stable IP. "NodePort" builds on ClusterIP and allocates a port on every node which routes to the clusterIP. "LoadBalancer" builds on NodePort and creates an external load-balancer (if supported in the current cloud) which routes to the clusterIP. More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceStatus.libsonnet new file mode 100644 index 00000000000..0c17a2e987d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/serviceStatus.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceStatus', url='', help='ServiceStatus represents the current status of a service.'), + '#loadBalancer':: d.obj(help='LoadBalancerStatus represents the status of a load-balancer.'), + loadBalancer: { + '#withIngress':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.', args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { loadBalancer+: { ingress: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withIngressMixin':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { loadBalancer+: { ingress+: if std.isArray(v=ingress) then ingress else [ingress] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sessionAffinityConfig.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sessionAffinityConfig.libsonnet new file mode 100644 index 00000000000..ca7e743ae4d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sessionAffinityConfig.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='sessionAffinityConfig', url='', help='SessionAffinityConfig represents the configurations of session affinity.'), + '#clientIP':: d.obj(help='ClientIPConfig represents the configurations of Client IP based session affinity.'), + clientIP: { + '#withTimeoutSeconds':: d.fn(help='timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours).', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { clientIP+: { timeoutSeconds: timeoutSeconds } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSPersistentVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSPersistentVolumeSource.libsonnet new file mode 100644 index 00000000000..ea679a2a4f4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSPersistentVolumeSource.libsonnet @@ -0,0 +1,31 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='storageOSPersistentVolumeSource', url='', help='Represents a StorageOS persistent volume resource.'), + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { secretRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { secretRef+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { secretRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { secretRef+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { secretRef+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { secretRef+: { uid: uid } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeName: volumeName }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { volumeNamespace: volumeNamespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSVolumeSource.libsonnet new file mode 100644 index 00000000000..a359b2268fe --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/storageOSVolumeSource.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='storageOSVolumeSource', url='', help='Represents a StorageOS persistent volume resource.'), + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { volumeName: volumeName }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { volumeNamespace: volumeNamespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sysctl.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sysctl.libsonnet new file mode 100644 index 00000000000..bf06bdc9830 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/sysctl.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='sysctl', url='', help='Sysctl defines a kernel parameter to be set'), + '#withName':: d.fn(help='Name of a property to set', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='Value of a property to set', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/taint.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/taint.libsonnet new file mode 100644 index 00000000000..5986eb0dd1d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/taint.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='taint', url='', help='The node this Taint is attached to has the "effect" on any pod that does not tolerate the Taint.'), + '#withEffect':: d.fn(help='Required. The effect of the taint on pods that do not tolerate the taint. Valid effects are NoSchedule, PreferNoSchedule and NoExecute.', args=[d.arg(name='effect', type=d.T.string)]), + withEffect(effect): { effect: effect }, + '#withKey':: d.fn(help='Required. The taint key to be applied to a node.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withTimeAdded':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='timeAdded', type=d.T.string)]), + withTimeAdded(timeAdded): { timeAdded: timeAdded }, + '#withValue':: d.fn(help='The taint value corresponding to the taint key.', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/tcpSocketAction.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/tcpSocketAction.libsonnet new file mode 100644 index 00000000000..af7af7889f7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/tcpSocketAction.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='tcpSocketAction', url='', help='TCPSocketAction describes an action based on opening a socket'), + '#withHost':: d.fn(help='Optional: Host name to connect to, defaults to the pod IP.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { port: port }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/toleration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/toleration.libsonnet new file mode 100644 index 00000000000..8804a8a65ff --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/toleration.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='toleration', url='', help='The pod this Toleration is attached to tolerates any taint that matches the triple using the matching operator .'), + '#withEffect':: d.fn(help='Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute.', args=[d.arg(name='effect', type=d.T.string)]), + withEffect(effect): { effect: effect }, + '#withKey':: d.fn(help='Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category.", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withTolerationSeconds':: d.fn(help='TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system.', args=[d.arg(name='tolerationSeconds', type=d.T.integer)]), + withTolerationSeconds(tolerationSeconds): { tolerationSeconds: tolerationSeconds }, + '#withValue':: d.fn(help='Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string.', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorLabelRequirement.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorLabelRequirement.libsonnet new file mode 100644 index 00000000000..35cb48a80f5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorLabelRequirement.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='topologySelectorLabelRequirement', url='', help='A topology selector requirement is a selector that matches given label. This is an alpha feature and may change in the future.'), + '#withKey':: d.fn(help='The label key that the selector applies to.', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withValues':: d.fn(help='An array of string values. One value must match the label to be selected. Each entry in Values is ORed.', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='An array of string values. One value must match the label to be selected. Each entry in Values is ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorTerm.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorTerm.libsonnet new file mode 100644 index 00000000000..6bf16c44d52 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySelectorTerm.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='topologySelectorTerm', url='', help='A topology selector term represents the result of label queries. A null or empty topology selector term matches no objects. The requirements of them are ANDed. It provides a subset of functionality as NodeSelectorTerm. This is an alpha feature and may change in the future.'), + '#withMatchLabelExpressions':: d.fn(help='A list of topology selector requirements by labels.', args=[d.arg(name='matchLabelExpressions', type=d.T.array)]), + withMatchLabelExpressions(matchLabelExpressions): { matchLabelExpressions: if std.isArray(v=matchLabelExpressions) then matchLabelExpressions else [matchLabelExpressions] }, + '#withMatchLabelExpressionsMixin':: d.fn(help='A list of topology selector requirements by labels.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabelExpressions', type=d.T.array)]), + withMatchLabelExpressionsMixin(matchLabelExpressions): { matchLabelExpressions+: if std.isArray(v=matchLabelExpressions) then matchLabelExpressions else [matchLabelExpressions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySpreadConstraint.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySpreadConstraint.libsonnet new file mode 100644 index 00000000000..68ffe773d26 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/topologySpreadConstraint.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='topologySpreadConstraint', url='', help='TopologySpreadConstraint specifies how to spread matching pods among the given topology.'), + '#labelSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + labelSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#withMaxSkew':: d.fn(help="MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 1/1/0: | zone1 | zone2 | zone3 | | P | P | | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 1/1/1; scheduling it onto zone1(zone2) would make the ActualSkew(2-0) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed.", args=[d.arg(name='maxSkew', type=d.T.integer)]), + withMaxSkew(maxSkew): { maxSkew: maxSkew }, + '#withTopologyKey':: d.fn(help="TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each as a 'bucket', and try to put balanced number of pods into each bucket. It's a required field.", args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + '#withWhenUnsatisfiable':: d.fn(help="WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location,\n but giving higher precedence to topologies that would help reduce the\n skew.\nA constraint is considered 'Unsatisfiable' for an incoming pod if and only if every possible node assigment for that pod would violate 'MaxSkew' on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field.", args=[d.arg(name='whenUnsatisfiable', type=d.T.string)]), + withWhenUnsatisfiable(whenUnsatisfiable): { whenUnsatisfiable: whenUnsatisfiable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/typedLocalObjectReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/typedLocalObjectReference.libsonnet new file mode 100644 index 00000000000..62c906f3ffd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/typedLocalObjectReference.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='typedLocalObjectReference', url='', help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volume.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volume.libsonnet new file mode 100644 index 00000000000..1704c748c6b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volume.libsonnet @@ -0,0 +1,475 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volume', url='', help='Volume represents a named volume in a pod that may be accessed by any container in the pod.'), + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { awsElasticBlockStore+: { fsType: fsType } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { awsElasticBlockStore+: { partition: partition } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { awsElasticBlockStore+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { awsElasticBlockStore+: { volumeID: volumeID } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { azureDisk+: { cachingMode: cachingMode } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { azureDisk+: { diskName: diskName } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { azureDisk+: { diskURI: diskURI } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { azureDisk+: { fsType: fsType } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { azureDisk+: { kind: kind } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureDisk+: { readOnly: readOnly } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureFile+: { readOnly: readOnly } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { azureFile+: { secretName: secretName } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { azureFile+: { shareName: shareName } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cephfs+: { secretRef+: { name: name } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { cephfs+: { path: path } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cephfs+: { readOnly: readOnly } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { cephfs+: { secretFile: secretFile } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { cephfs+: { user: user } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cinder+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { cinder+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cinder+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { cinder+: { volumeID: volumeID } }, + }, + '#configMap':: d.obj(help="Adapts a ConfigMap into a volume.\n\nThe contents of the target ConfigMap's Data field will be presented in a volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. ConfigMap volumes support ownership management and SELinux relabeling."), + configMap: { + '#withDefaultMode':: d.fn(help='Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { configMap+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#csi':: d.obj(help='Represents a source location of a volume to mount, managed by an external CSI driver'), + csi: { + '#nodePublishSecretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { nodePublishSecretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { csi+: { driver: driver } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Ex. "ext4", "xfs", "ntfs". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { csi+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Specifies a read-only configuration for the volume. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { csi+: { readOnly: readOnly } }, + '#withVolumeAttributes':: d.fn(help="VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { csi+: { volumeAttributes: volumeAttributes } }, + '#withVolumeAttributesMixin':: d.fn(help="VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { csi+: { volumeAttributes+: volumeAttributes } }, + }, + '#downwardAPI':: d.obj(help='DownwardAPIVolumeSource represents a volume containing downward API info. Downward API volumes support ownership management and SELinux relabeling.'), + downwardAPI: { + '#withDefaultMode':: d.fn(help='Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { downwardAPI+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help='Items is a list of downward API volume file', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='Items is a list of downward API volume file\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#emptyDir':: d.obj(help='Represents an empty directory for a pod. Empty directory volumes support ownership management and SELinux relabeling.'), + emptyDir: { + '#withMedium':: d.fn(help="What type of storage medium should back this directory. The default is '' which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { emptyDir+: { medium: medium } }, + '#withSizeLimit':: d.fn(help="Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and AsInt64() accessors.\n\nThe serialization format is:\n\n ::= \n (Note that may be empty, from the '' case in .)\n ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= '+' | '-' ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei\n (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html)\n ::= m | '' | k | M | G | T | P | E\n (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.)\n ::= 'e' | 'E' \n\nNo matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities.\n\nWhen a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized.\n\nBefore serializing, Quantity will be put in 'canonical form'. This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that:\n a. No precision is lost\n b. No fractional digits will be emitted\n c. The exponent (or suffix) is as large as possible.\nThe sign will be omitted unless the number is negative.\n\nExamples:\n 1.5 will be serialized as '1500m'\n 1.5Gi will be serialized as '1536Mi'\n\nNote that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise.\n\nNon-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.)\n\nThis format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation.", args=[d.arg(name='sizeLimit', type=d.T.string)]), + withSizeLimit(sizeLimit): { emptyDir+: { sizeLimit: sizeLimit } }, + }, + '#ephemeral':: d.obj(help='Represents an ephemeral volume that is handled by a normal storage driver.'), + ephemeral: { + '#volumeClaimTemplate':: d.obj(help='PersistentVolumeClaimTemplate is used to produce PersistentVolumeClaim objects as part of an EphemeralVolumeSource.'), + volumeClaimTemplate: { + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { ephemeral+: { volumeClaimTemplate+: { metadata+: { annotations: annotations } } } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { ephemeral+: { volumeClaimTemplate+: { metadata+: { annotations+: annotations } } } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { ephemeral+: { volumeClaimTemplate+: { metadata+: { clusterName: clusterName } } } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { ephemeral+: { volumeClaimTemplate+: { metadata+: { creationTimestamp: creationTimestamp } } } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { ephemeral+: { volumeClaimTemplate+: { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } } } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { ephemeral+: { volumeClaimTemplate+: { metadata+: { deletionTimestamp: deletionTimestamp } } } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { ephemeral+: { volumeClaimTemplate+: { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { ephemeral+: { volumeClaimTemplate+: { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } } } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { ephemeral+: { volumeClaimTemplate+: { metadata+: { generateName: generateName } } } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { ephemeral+: { volumeClaimTemplate+: { metadata+: { generation: generation } } } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { ephemeral+: { volumeClaimTemplate+: { metadata+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { ephemeral+: { volumeClaimTemplate+: { metadata+: { labels+: labels } } } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { ephemeral+: { volumeClaimTemplate+: { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { ephemeral+: { volumeClaimTemplate+: { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } } } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { metadata+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { ephemeral+: { volumeClaimTemplate+: { metadata+: { namespace: namespace } } } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { ephemeral+: { volumeClaimTemplate+: { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { ephemeral+: { volumeClaimTemplate+: { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } } } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { ephemeral+: { volumeClaimTemplate+: { metadata+: { resourceVersion: resourceVersion } } } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { ephemeral+: { volumeClaimTemplate+: { metadata+: { selfLink: selfLink } } } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { ephemeral+: { volumeClaimTemplate+: { metadata+: { uid: uid } } } }, + }, + '#spec':: d.obj(help='PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes'), + spec: { + '#dataSource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + dataSource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } }, + }, + '#resources':: d.obj(help='ResourceRequirements describes the compute resource requirements.'), + resources: { + '#withLimits':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } }, + '#withLimitsMixin':: d.fn(help='Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } }, + '#withRequests':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } }, + '#withRequestsMixin':: d.fn(help='Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } }, + }, + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withStorageClassName':: d.fn(help='Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the binding reference to the PersistentVolume backing this claim.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } }, + }, + }, + '#withReadOnly':: d.fn(help='Specifies a read-only configuration for the volume. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { ephemeral+: { readOnly: readOnly } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fc+: { fsType: fsType } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { fc+: { lun: lun } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { fc+: { readOnly: readOnly } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } }, + }, + '#flexVolume':: d.obj(help='FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { flexVolume+: { secretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { flexVolume+: { driver: driver } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { flexVolume+: { fsType: fsType } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { flexVolume+: { options: options } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { flexVolume+: { options+: options } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { flexVolume+: { readOnly: readOnly } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { flocker+: { datasetName: datasetName } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { flocker+: { datasetUUID: datasetUUID } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { gcePersistentDisk+: { fsType: fsType } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { gcePersistentDisk+: { partition: partition } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { gcePersistentDisk+: { pdName: pdName } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { gcePersistentDisk+: { readOnly: readOnly } }, + }, + '#gitRepo':: d.obj(help="Represents a volume that is populated with the contents of a git repository. Git repo volumes do not support ownership management. Git repo volumes support SELinux relabeling.\n\nDEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container."), + gitRepo: { + '#withDirectory':: d.fn(help="Target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name.", args=[d.arg(name='directory', type=d.T.string)]), + withDirectory(directory): { gitRepo+: { directory: directory } }, + '#withRepository':: d.fn(help='Repository URL', args=[d.arg(name='repository', type=d.T.string)]), + withRepository(repository): { gitRepo+: { repository: repository } }, + '#withRevision':: d.fn(help='Commit hash for the specified revision.', args=[d.arg(name='revision', type=d.T.string)]), + withRevision(revision): { gitRepo+: { revision: revision } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { glusterfs+: { endpoints: endpoints } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { glusterfs+: { path: path } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { glusterfs+: { readOnly: readOnly } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { hostPath+: { path: path } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { hostPath+: { type: type } }, + }, + '#iscsi':: d.obj(help='Represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { iscsi+: { secretRef+: { name: name } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { iscsi+: { chapAuthSession: chapAuthSession } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { iscsi+: { fsType: fsType } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { iscsi+: { initiatorName: initiatorName } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iscsi+: { iqn: iqn } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsi+: { iscsiInterface: iscsiInterface } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { iscsi+: { lun: lun } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { iscsi+: { readOnly: readOnly } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { iscsi+: { targetPortal: targetPortal } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nfs+: { path: path } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { nfs+: { readOnly: readOnly } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { nfs+: { server: server } }, + }, + '#persistentVolumeClaim':: d.obj(help="PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. This volume finds the bound PV and mounts that volume for the pod. A PersistentVolumeClaimVolumeSource is, essentially, a wrapper around another type of volume that is owned by someone else (the system)."), + persistentVolumeClaim: { + '#withClaimName':: d.fn(help='ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims', args=[d.arg(name='claimName', type=d.T.string)]), + withClaimName(claimName): { persistentVolumeClaim+: { claimName: claimName } }, + '#withReadOnly':: d.fn(help='Will force the ReadOnly setting in VolumeMounts. Default false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { persistentVolumeClaim+: { readOnly: readOnly } }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { photonPersistentDisk+: { fsType: fsType } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { photonPersistentDisk+: { pdID: pdID } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { portworxVolume+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { portworxVolume+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { portworxVolume+: { volumeID: volumeID } }, + }, + '#projected':: d.obj(help='Represents a projected volume source'), + projected: { + '#withDefaultMode':: d.fn(help='Mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { projected+: { defaultMode: defaultMode } }, + '#withSources':: d.fn(help='list of volume projections', args=[d.arg(name='sources', type=d.T.array)]), + withSources(sources): { projected+: { sources: if std.isArray(v=sources) then sources else [sources] } }, + '#withSourcesMixin':: d.fn(help='list of volume projections\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]), + withSourcesMixin(sources): { projected+: { sources+: if std.isArray(v=sources) then sources else [sources] } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { quobyte+: { group: group } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { quobyte+: { readOnly: readOnly } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { quobyte+: { registry: registry } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { quobyte+: { tenant: tenant } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { quobyte+: { user: user } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { quobyte+: { volume: volume } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { rbd+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { rbd+: { fsType: fsType } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { rbd+: { image: image } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { rbd+: { keyring: keyring } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { rbd+: { pool: pool } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { rbd+: { readOnly: readOnly } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { rbd+: { user: user } }, + }, + '#scaleIO':: d.obj(help='ScaleIOVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleIO+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { scaleIO+: { fsType: fsType } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { scaleIO+: { gateway: gateway } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { scaleIO+: { protectionDomain: protectionDomain } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { scaleIO+: { readOnly: readOnly } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { scaleIO+: { sslEnabled: sslEnabled } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { scaleIO+: { storageMode: storageMode } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { scaleIO+: { storagePool: storagePool } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { scaleIO+: { system: system } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { scaleIO+: { volumeName: volumeName } }, + }, + '#secret':: d.obj(help="Adapts a Secret into a volume.\n\nThe contents of the target Secret's Data field will be presented in a volume as files using the keys in the Data field as the file names. Secret volumes support ownership management and SELinux relabeling."), + secret: { + '#withDefaultMode':: d.fn(help='Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set.', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { secret+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withOptional':: d.fn(help='Specify whether the Secret or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + '#withSecretName':: d.fn(help="Name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret", args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secret+: { secretName: secretName } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace.'), + secretRef: { + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { storageos+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { storageos+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { storageos+: { readOnly: readOnly } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { storageos+: { volumeName: volumeName } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { storageos+: { volumeNamespace: volumeNamespace } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { vsphereVolume+: { fsType: fsType } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { vsphereVolume+: { storagePolicyID: storagePolicyID } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { vsphereVolume+: { storagePolicyName: storagePolicyName } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { vsphereVolume+: { volumePath: volumePath } }, + }, + '#withName':: d.fn(help="Volume's name. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names", args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeDevice.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeDevice.libsonnet new file mode 100644 index 00000000000..31e95a656e9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeDevice.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeDevice', url='', help='volumeDevice describes a mapping of a raw block device within a container.'), + '#withDevicePath':: d.fn(help='devicePath is the path inside of the container that the device will be mapped to.', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='name must match the name of a persistentVolumeClaim in the pod', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeMount.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeMount.libsonnet new file mode 100644 index 00000000000..af137c887e3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeMount.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeMount', url='', help='VolumeMount describes a mounting of a Volume within a container.'), + '#withMountPath':: d.fn(help="Path within the container at which the volume should be mounted. Must not contain ':'.", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10.', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='This must match the Name of a Volume.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="Path within the volume from which the container's volume should be mounted. Defaults to '' (volume's root).", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to '' (volume's root). SubPathExpr and SubPath are mutually exclusive.", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeNodeAffinity.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeNodeAffinity.libsonnet new file mode 100644 index 00000000000..d57b6de795a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeNodeAffinity.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeNodeAffinity', url='', help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeProjection.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeProjection.libsonnet new file mode 100644 index 00000000000..d517689d234 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/volumeProjection.libsonnet @@ -0,0 +1,44 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeProjection', url='', help='Projection that may be projected along with other supported volume types'), + '#configMap':: d.obj(help="Adapts a ConfigMap into a projected volume.\n\nThe contents of the target ConfigMap's Data field will be presented in a projected volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. Note that this is identical to a configmap volume source without the default mode."), + configMap: { + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the ConfigMap or its keys must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#downwardAPI':: d.obj(help='Represents downward API info for projecting into a projected volume. Note that this is identical to a downwardAPI volume source without the default mode.'), + downwardAPI: { + '#withItems':: d.fn(help='Items is a list of DownwardAPIVolume file', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='Items is a list of DownwardAPIVolume file\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#secret':: d.obj(help="Adapts a secret into a projected volume.\n\nThe contents of the target Secret's Data field will be presented in a projected volume as files using the keys in the Data field as the file names. Note that this is identical to a secret volume source without the default mode."), + secret: { + '#withItems':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secret+: { name: name } }, + '#withOptional':: d.fn(help='Specify whether the Secret or its key must be defined', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + }, + '#serviceAccountToken':: d.obj(help='ServiceAccountTokenProjection represents a projected service account token volume. This projection can be used to insert a service account token into the pods runtime filesystem for use against APIs (Kubernetes API Server or otherwise).'), + serviceAccountToken: { + '#withAudience':: d.fn(help='Audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver.', args=[d.arg(name='audience', type=d.T.string)]), + withAudience(audience): { serviceAccountToken+: { audience: audience } }, + '#withExpirationSeconds':: d.fn(help='ExpirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes.', args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { serviceAccountToken+: { expirationSeconds: expirationSeconds } }, + '#withPath':: d.fn(help='Path is the path relative to the mount point of the file to project the token into.', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { serviceAccountToken+: { path: path } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/vsphereVirtualDiskVolumeSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/vsphereVirtualDiskVolumeSource.libsonnet new file mode 100644 index 00000000000..baa77e0e166 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/vsphereVirtualDiskVolumeSource.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='vsphereVirtualDiskVolumeSource', url='', help='Represents a vSphere volume resource.'), + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fsType: fsType }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { storagePolicyID: storagePolicyID }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { storagePolicyName: storagePolicyName }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { volumePath: volumePath }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/weightedPodAffinityTerm.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/weightedPodAffinityTerm.libsonnet new file mode 100644 index 00000000000..0b4eb298bbb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/weightedPodAffinityTerm.libsonnet @@ -0,0 +1,28 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='weightedPodAffinityTerm', url='', help='The weights of all of the matched WeightedPodAffinityTerm fields are added per-node to find the most preferred node(s)'), + '#podAffinityTerm':: d.obj(help='Defines a set of pods (namely those matching the labelSelector relative to the given namespace(s)) that this pod should be co-located (affinity) or not co-located (anti-affinity) with, where co-located is defined as running on a node whose value of the label with key matches that of any node on which a pod of the set of pods is running'), + podAffinityTerm: { + '#labelSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + labelSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } }, + }, + '#withNamespaces':: d.fn(help="namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means 'this pod's namespace'", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withNamespacesMixin':: d.fn(help="namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means 'this pod's namespace'\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withTopologyKey':: d.fn(help='This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed.', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } }, + }, + '#withWeight':: d.fn(help='weight associated with matching the corresponding podAffinityTerm, in the range 1-100.', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/windowsSecurityContextOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/windowsSecurityContextOptions.libsonnet new file mode 100644 index 00000000000..ee6426faa20 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/core/v1/windowsSecurityContextOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='windowsSecurityContextOptions', url='', help='WindowsSecurityContextOptions contain Windows-specific options and credentials.'), + '#withGmsaCredentialSpec':: d.fn(help='GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field.', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { gmsaCredentialSpec: gmsaCredentialSpec }, + '#withGmsaCredentialSpecName':: d.fn(help='GMSACredentialSpecName is the name of the GMSA credential spec to use.', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { gmsaCredentialSpecName: gmsaCredentialSpecName }, + '#withRunAsUserName':: d.fn(help='The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence.', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { runAsUserName: runAsUserName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/main.libsonnet new file mode 100644 index 00000000000..bc83e7bed9b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='discovery', url='', help=''), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpoint.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpoint.libsonnet new file mode 100644 index 00000000000..8bc9fd5cda3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpoint.libsonnet @@ -0,0 +1,38 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpoint', url='', help='Endpoint represents a single logical "backend" implementing a service.'), + '#conditions':: d.obj(help='EndpointConditions represents the current condition of an endpoint.'), + conditions: { + '#withReady':: d.fn(help='ready indicates that this endpoint is prepared to receive traffic, according to whatever system is managing the endpoint. A nil value indicates an unknown state. In most cases consumers should interpret this unknown state as ready.', args=[d.arg(name='ready', type=d.T.boolean)]), + withReady(ready): { conditions+: { ready: ready } }, + }, + '#targetRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + targetRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { targetRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { targetRef+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { targetRef+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { targetRef+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { targetRef+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { targetRef+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { targetRef+: { uid: uid } }, + }, + '#withAddresses':: d.fn(help='addresses of this endpoint. The contents of this field are interpreted according to the corresponding EndpointSlice addressType field. Consumers must handle different types of addresses in the context of their own capabilities. This must contain at least one address but no more than 100.', args=[d.arg(name='addresses', type=d.T.array)]), + withAddresses(addresses): { addresses: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withAddressesMixin':: d.fn(help='addresses of this endpoint. The contents of this field are interpreted according to the corresponding EndpointSlice addressType field. Consumers must handle different types of addresses in the context of their own capabilities. This must contain at least one address but no more than 100.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='addresses', type=d.T.array)]), + withAddressesMixin(addresses): { addresses+: if std.isArray(v=addresses) then addresses else [addresses] }, + '#withHostname':: d.fn(help='hostname of this endpoint. This field may be used by consumers of endpoints to distinguish endpoints from each other (e.g. in DNS names). Multiple endpoints which use the same hostname should be considered fungible (e.g. multiple A values in DNS). Must pass DNS Label (RFC 1123) validation.', args=[d.arg(name='hostname', type=d.T.string)]), + withHostname(hostname): { hostname: hostname }, + '#withTopology':: d.fn(help='topology contains arbitrary topology information associated with the endpoint. These key/value pairs must conform with the label format. https://kubernetes.io/docs/concepts/overview/working-with-objects/labels Topology may include a maximum of 16 key/value pairs. This includes, but is not limited to the following well known keys: * kubernetes.io/hostname: the value indicates the hostname of the node\n where the endpoint is located. This should match the corresponding\n node label.\n* topology.kubernetes.io/zone: the value indicates the zone where the\n endpoint is located. This should match the corresponding node label.\n* topology.kubernetes.io/region: the value indicates the region where the\n endpoint is located. This should match the corresponding node label.', args=[d.arg(name='topology', type=d.T.object)]), + withTopology(topology): { topology: topology }, + '#withTopologyMixin':: d.fn(help='topology contains arbitrary topology information associated with the endpoint. These key/value pairs must conform with the label format. https://kubernetes.io/docs/concepts/overview/working-with-objects/labels Topology may include a maximum of 16 key/value pairs. This includes, but is not limited to the following well known keys: * kubernetes.io/hostname: the value indicates the hostname of the node\n where the endpoint is located. This should match the corresponding\n node label.\n* topology.kubernetes.io/zone: the value indicates the zone where the\n endpoint is located. This should match the corresponding node label.\n* topology.kubernetes.io/region: the value indicates the region where the\n endpoint is located. This should match the corresponding node label.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topology', type=d.T.object)]), + withTopologyMixin(topology): { topology+: topology }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointConditions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointConditions.libsonnet new file mode 100644 index 00000000000..2d8baf972eb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointConditions.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointConditions', url='', help='EndpointConditions represents the current condition of an endpoint.'), + '#withReady':: d.fn(help='ready indicates that this endpoint is prepared to receive traffic, according to whatever system is managing the endpoint. A nil value indicates an unknown state. In most cases consumers should interpret this unknown state as ready.', args=[d.arg(name='ready', type=d.T.boolean)]), + withReady(ready): { ready: ready }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointPort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointPort.libsonnet new file mode 100644 index 00000000000..d127c6c7940 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointPort.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointPort', url='', help='EndpointPort represents a Port used by an EndpointSlice'), + '#withAppProtocol':: d.fn(help='The application protocol for this port. This field follows standard Kubernetes label syntax. Un-prefixed names are reserved for IANA standard service names (as per RFC-6335 and http://www.iana.org/assignments/service-names). Non-standard protocols should use prefixed names such as mycompany.com/my-custom-protocol.', args=[d.arg(name='appProtocol', type=d.T.string)]), + withAppProtocol(appProtocol): { appProtocol: appProtocol }, + '#withName':: d.fn(help="The name of this port. All ports in an EndpointSlice must have a unique name. If the EndpointSlice is dervied from a Kubernetes service, this corresponds to the Service.ports[].name. Name must either be an empty string or pass DNS_LABEL validation: * must be no more than 63 characters long. * must consist of lower case alphanumeric characters or '-'. * must start and end with an alphanumeric character. Default is empty string.", args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPort':: d.fn(help='The port number of the endpoint. If this is not specified, ports are not restricted and must be interpreted in the context of the specific consumer.', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { port: port }, + '#withProtocol':: d.fn(help='The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP.', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointSlice.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointSlice.libsonnet new file mode 100644 index 00000000000..75844b796c0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/endpointSlice.libsonnet @@ -0,0 +1,66 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='endpointSlice', url='', help='EndpointSlice represents a subset of the endpoints that implement a service. For a given service there may be multiple EndpointSlice objects, selected by labels, which must be joined to produce the full set of endpoints.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Endpointslice', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'discovery.k8s.io/v1beta1', + kind: 'EndpointSlice', + } + self.metadata.withName(name=name), + '#withAddressType':: d.fn(help='addressType specifies the type of address carried by this EndpointSlice. All addresses in this slice must be the same type. This field is immutable after creation. The following address types are currently supported: * IPv4: Represents an IPv4 Address. * IPv6: Represents an IPv6 Address. * FQDN: Represents a Fully Qualified Domain Name.', args=[d.arg(name='addressType', type=d.T.string)]), + withAddressType(addressType): { addressType: addressType }, + '#withEndpoints':: d.fn(help='endpoints is a list of unique endpoints in this slice. Each slice may include a maximum of 1000 endpoints.', args=[d.arg(name='endpoints', type=d.T.array)]), + withEndpoints(endpoints): { endpoints: if std.isArray(v=endpoints) then endpoints else [endpoints] }, + '#withEndpointsMixin':: d.fn(help='endpoints is a list of unique endpoints in this slice. Each slice may include a maximum of 1000 endpoints.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpoints', type=d.T.array)]), + withEndpointsMixin(endpoints): { endpoints+: if std.isArray(v=endpoints) then endpoints else [endpoints] }, + '#withPorts':: d.fn(help='ports specifies the list of network ports exposed by each endpoint in this slice. Each port must have a unique name. When ports is empty, it indicates that there are no defined ports. When a port is defined with a nil port value, it indicates "all ports". Each slice may include a maximum of 100 ports.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='ports specifies the list of network ports exposed by each endpoint in this slice. Each port must have a unique name. When ports is empty, it indicates that there are no defined ports. When a port is defined with a nil port value, it indicates "all ports". Each slice may include a maximum of 100 ports.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/main.libsonnet new file mode 100644 index 00000000000..408374e6e0f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/discovery/v1beta1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + endpoint: (import 'endpoint.libsonnet'), + endpointConditions: (import 'endpointConditions.libsonnet'), + endpointPort: (import 'endpointPort.libsonnet'), + endpointSlice: (import 'endpointSlice.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/main.libsonnet new file mode 100644 index 00000000000..c60a9f4d15d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='events', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/event.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/event.libsonnet new file mode 100644 index 00000000000..eb503edff77 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/event.libsonnet @@ -0,0 +1,124 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='event', url='', help='Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system.'), + '#deprecatedSource':: d.obj(help='EventSource contains information for an event.'), + deprecatedSource: { + '#withComponent':: d.fn(help='Component from which the event is generated.', args=[d.arg(name='component', type=d.T.string)]), + withComponent(component): { deprecatedSource+: { component: component } }, + '#withHost':: d.fn(help='Node name on which the event is generated.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { deprecatedSource+: { host: host } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Event', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'events.k8s.io/v1', + kind: 'Event', + } + self.metadata.withName(name=name), + '#regarding':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + regarding: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { regarding+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { regarding+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { regarding+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { regarding+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { regarding+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { regarding+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { regarding+: { uid: uid } }, + }, + '#related':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + related: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { related+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { related+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { related+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { related+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { related+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { related+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { related+: { uid: uid } }, + }, + '#series':: d.obj(help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time. How often to update the EventSeries is up to the event reporters. The default event reporter in "k8s.io/client-go/tools/events/event_broadcaster.go" shows how this struct is updated on heartbeats and can guide customized reporter implementations.'), + series: { + '#withCount':: d.fn(help='count is the number of occurrences in this series up to the last heartbeat time.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { series+: { count: count } }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { series+: { lastObservedTime: lastObservedTime } }, + }, + '#withAction':: d.fn(help='action is what action was taken/failed regarding to the regarding object. It is machine-readable. This field can have at most 128 characters.', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withDeprecatedCount':: d.fn(help='deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type.', args=[d.arg(name='deprecatedCount', type=d.T.integer)]), + withDeprecatedCount(deprecatedCount): { deprecatedCount: deprecatedCount }, + '#withDeprecatedFirstTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deprecatedFirstTimestamp', type=d.T.string)]), + withDeprecatedFirstTimestamp(deprecatedFirstTimestamp): { deprecatedFirstTimestamp: deprecatedFirstTimestamp }, + '#withDeprecatedLastTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deprecatedLastTimestamp', type=d.T.string)]), + withDeprecatedLastTimestamp(deprecatedLastTimestamp): { deprecatedLastTimestamp: deprecatedLastTimestamp }, + '#withEventTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='eventTime', type=d.T.string)]), + withEventTime(eventTime): { eventTime: eventTime }, + '#withNote':: d.fn(help='note is a human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB.', args=[d.arg(name='note', type=d.T.string)]), + withNote(note): { note: note }, + '#withReason':: d.fn(help='reason is why the action was taken. It is human-readable. This field can have at most 128 characters.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withReportingController':: d.fn(help='reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. This field cannot be empty for new Events.', args=[d.arg(name='reportingController', type=d.T.string)]), + withReportingController(reportingController): { reportingController: reportingController }, + '#withReportingInstance':: d.fn(help='reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. This field cannot be empty for new Events and it can have at most 128 characters.', args=[d.arg(name='reportingInstance', type=d.T.string)]), + withReportingInstance(reportingInstance): { reportingInstance: reportingInstance }, + '#withType':: d.fn(help='type is the type of this event (Normal, Warning), new types could be added in the future. It is machine-readable.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/eventSeries.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/eventSeries.libsonnet new file mode 100644 index 00000000000..a88c1e1734c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/eventSeries.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='eventSeries', url='', help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time. How often to update the EventSeries is up to the event reporters. The default event reporter in "k8s.io/client-go/tools/events/event_broadcaster.go" shows how this struct is updated on heartbeats and can guide customized reporter implementations.'), + '#withCount':: d.fn(help='count is the number of occurrences in this series up to the last heartbeat time.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { lastObservedTime: lastObservedTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/main.libsonnet new file mode 100644 index 00000000000..8e90a2e00c4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + event: (import 'event.libsonnet'), + eventSeries: (import 'eventSeries.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/event.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/event.libsonnet new file mode 100644 index 00000000000..77b1d8802b1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/event.libsonnet @@ -0,0 +1,124 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='event', url='', help='Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system.'), + '#deprecatedSource':: d.obj(help='EventSource contains information for an event.'), + deprecatedSource: { + '#withComponent':: d.fn(help='Component from which the event is generated.', args=[d.arg(name='component', type=d.T.string)]), + withComponent(component): { deprecatedSource+: { component: component } }, + '#withHost':: d.fn(help='Node name on which the event is generated.', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { deprecatedSource+: { host: host } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Event', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'events.k8s.io/v1beta1', + kind: 'Event', + } + self.metadata.withName(name=name), + '#regarding':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + regarding: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { regarding+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { regarding+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { regarding+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { regarding+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { regarding+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { regarding+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { regarding+: { uid: uid } }, + }, + '#related':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + related: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { related+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { related+: { fieldPath: fieldPath } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { related+: { kind: kind } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { related+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { related+: { namespace: namespace } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { related+: { resourceVersion: resourceVersion } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { related+: { uid: uid } }, + }, + '#series':: d.obj(help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.'), + series: { + '#withCount':: d.fn(help='count is the number of occurrences in this series up to the last heartbeat time.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { series+: { count: count } }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { series+: { lastObservedTime: lastObservedTime } }, + }, + '#withAction':: d.fn(help='action is what action was taken/failed regarding to the regarding object. It is machine-readable. This field can have at most 128 characters.', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withDeprecatedCount':: d.fn(help='deprecatedCount is the deprecated field assuring backward compatibility with core.v1 Event type.', args=[d.arg(name='deprecatedCount', type=d.T.integer)]), + withDeprecatedCount(deprecatedCount): { deprecatedCount: deprecatedCount }, + '#withDeprecatedFirstTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deprecatedFirstTimestamp', type=d.T.string)]), + withDeprecatedFirstTimestamp(deprecatedFirstTimestamp): { deprecatedFirstTimestamp: deprecatedFirstTimestamp }, + '#withDeprecatedLastTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deprecatedLastTimestamp', type=d.T.string)]), + withDeprecatedLastTimestamp(deprecatedLastTimestamp): { deprecatedLastTimestamp: deprecatedLastTimestamp }, + '#withEventTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='eventTime', type=d.T.string)]), + withEventTime(eventTime): { eventTime: eventTime }, + '#withNote':: d.fn(help='note is a human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB.', args=[d.arg(name='note', type=d.T.string)]), + withNote(note): { note: note }, + '#withReason':: d.fn(help='reason is why the action was taken. It is human-readable. This field can have at most 128 characters.', args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withReportingController':: d.fn(help='reportingController is the name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. This field cannot be empty for new Events.', args=[d.arg(name='reportingController', type=d.T.string)]), + withReportingController(reportingController): { reportingController: reportingController }, + '#withReportingInstance':: d.fn(help='reportingInstance is the ID of the controller instance, e.g. `kubelet-xyzf`. This field cannot be empty for new Events and it can have at most 128 characters.', args=[d.arg(name='reportingInstance', type=d.T.string)]), + withReportingInstance(reportingInstance): { reportingInstance: reportingInstance }, + '#withType':: d.fn(help='type is the type of this event (Normal, Warning), new types could be added in the future. It is machine-readable.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/eventSeries.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/eventSeries.libsonnet new file mode 100644 index 00000000000..dd1a98ccaf9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/eventSeries.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='eventSeries', url='', help='EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time.'), + '#withCount':: d.fn(help='count is the number of occurrences in this series up to the last heartbeat time.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#withLastObservedTime':: d.fn(help='MicroTime is version of Time with microsecond level precision.', args=[d.arg(name='lastObservedTime', type=d.T.string)]), + withLastObservedTime(lastObservedTime): { lastObservedTime: lastObservedTime }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/main.libsonnet new file mode 100644 index 00000000000..34c50897a1c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/events/v1beta1/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + event: (import 'event.libsonnet'), + eventSeries: (import 'eventSeries.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/main.libsonnet new file mode 100644 index 00000000000..d67bd0cf44e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='extensions', url='', help=''), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressPath.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressPath.libsonnet new file mode 100644 index 00000000000..75e77b0c68d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressPath.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressPath', url='', help='HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.'), + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { backend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { backend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { resource+: { name: name } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { backend+: { serviceName: serviceName } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { backend+: { servicePort: servicePort } }, + }, + '#withPath':: d.fn(help="Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional 'path' part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPathType':: d.fn(help="PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types. Defaults to ImplementationSpecific.", args=[d.arg(name='pathType', type=d.T.string)]), + withPathType(pathType): { pathType: pathType }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressRuleValue.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressRuleValue.libsonnet new file mode 100644 index 00000000000..31169937a55 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/httpIngressRuleValue.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressRuleValue', url='', help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { paths: if std.isArray(v=paths) then paths else [paths] }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { paths+: if std.isArray(v=paths) then paths else [paths] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingress.libsonnet new file mode 100644 index 00000000000..680787571a4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingress.libsonnet @@ -0,0 +1,85 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingress', url='', help='Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc. DEPRECATED - This group version of Ingress is deprecated by networking.k8s.io/v1beta1 Ingress. See the release notes for more information.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Ingress', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'extensions/v1beta1', + kind: 'Ingress', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='IngressSpec describes the Ingress the user wishes to exist.'), + spec: { + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { backend+: { resource+: { apiGroup: apiGroup } } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { backend+: { resource+: { kind: kind } } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { backend+: { resource+: { name: name } } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { spec+: { backend+: { serviceName: serviceName } } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { spec+: { backend+: { servicePort: servicePort } } }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { spec+: { ingressClassName: ingressClassName } }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { spec+: { rules: if std.isArray(v=rules) then rules else [rules] } }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { spec+: { rules+: if std.isArray(v=rules) then rules else [rules] } }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { spec+: { tls: if std.isArray(v=tls) then tls else [tls] } }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { spec+: { tls+: if std.isArray(v=tls) then tls else [tls] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressBackend.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressBackend.libsonnet new file mode 100644 index 00000000000..1106f316187 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressBackend.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressBackend', url='', help='IngressBackend describes all endpoints for a given service and port.'), + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { resource+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { resource+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { serviceName: serviceName }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { servicePort: servicePort }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressRule.libsonnet new file mode 100644 index 00000000000..3a0a5dca509 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressRule.libsonnet @@ -0,0 +1,15 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressRule', url='', help='IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.'), + '#http':: d.obj(help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + http: { + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { http+: { paths: if std.isArray(v=paths) then paths else [paths] } }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { http+: { paths+: if std.isArray(v=paths) then paths else [paths] } }, + }, + '#withHost':: d.fn(help="Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the 'host' part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be 'precise' which is a domain name without the terminating dot of a network host (e.g. 'foo.bar.com') or 'wildcard', which is a domain name prefixed with a single wildcard label (e.g. '*.foo.com'). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == '*'). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressSpec.libsonnet new file mode 100644 index 00000000000..a61df751c89 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressSpec.libsonnet @@ -0,0 +1,32 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressSpec', url='', help='IngressSpec describes the Ingress the user wishes to exist.'), + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { backend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { backend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { resource+: { name: name } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { backend+: { serviceName: serviceName } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { backend+: { servicePort: servicePort } }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { ingressClassName: ingressClassName }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { tls: if std.isArray(v=tls) then tls else [tls] }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { tls+: if std.isArray(v=tls) then tls else [tls] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressStatus.libsonnet new file mode 100644 index 00000000000..697b5b33c1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressStatus.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressStatus', url='', help='IngressStatus describe the current state of the Ingress.'), + '#loadBalancer':: d.obj(help='LoadBalancerStatus represents the status of a load-balancer.'), + loadBalancer: { + '#withIngress':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.', args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { loadBalancer+: { ingress: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withIngressMixin':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { loadBalancer+: { ingress+: if std.isArray(v=ingress) then ingress else [ingress] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressTLS.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressTLS.libsonnet new file mode 100644 index 00000000000..0cddf5a4e45 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/ingressTLS.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressTLS', url='', help='IngressTLS describes the transport layer security associated with an Ingress.'), + '#withHosts':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.', args=[d.arg(name='hosts', type=d.T.array)]), + withHosts(hosts): { hosts: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withHostsMixin':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hosts', type=d.T.array)]), + withHostsMixin(hosts): { hosts+: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withSecretName':: d.fn(help='SecretName is the name of the secret used to terminate SSL traffic on 443. Field is left optional to allow SSL routing based on SNI hostname alone. If the SNI host in a listener conflicts with the "Host" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/main.libsonnet new file mode 100644 index 00000000000..c12f3f9af7c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/extensions/v1beta1/main.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + httpIngressPath: (import 'httpIngressPath.libsonnet'), + httpIngressRuleValue: (import 'httpIngressRuleValue.libsonnet'), + ingress: (import 'ingress.libsonnet'), + ingressBackend: (import 'ingressBackend.libsonnet'), + ingressRule: (import 'ingressRule.libsonnet'), + ingressSpec: (import 'ingressSpec.libsonnet'), + ingressStatus: (import 'ingressStatus.libsonnet'), + ingressTLS: (import 'ingressTLS.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/main.libsonnet new file mode 100644 index 00000000000..69897ddeab0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowcontrol', url='', help=''), + v1alpha1: (import 'v1alpha1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowDistinguisherMethod.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowDistinguisherMethod.libsonnet new file mode 100644 index 00000000000..649e1f633c3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowDistinguisherMethod.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowDistinguisherMethod', url='', help='FlowDistinguisherMethod specifies the method of a flow distinguisher.'), + '#withType':: d.fn(help='`type` is the type of flow distinguisher method The supported types are "ByUser" and "ByNamespace". Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchema.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchema.libsonnet new file mode 100644 index 00000000000..99e191b920c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchema.libsonnet @@ -0,0 +1,75 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowSchema', url='', help='FlowSchema defines the schema of a group of flows. Note that a flow is made up of a set of inbound API requests with similar attributes and is identified by a pair of strings: the name of the FlowSchema and a "flow distinguisher".'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Flowschema', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'flowcontrol.apiserver.k8s.io/v1alpha1', + kind: 'FlowSchema', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help="FlowSchemaSpec describes how the FlowSchema's specification looks like."), + spec: { + '#distinguisherMethod':: d.obj(help='FlowDistinguisherMethod specifies the method of a flow distinguisher.'), + distinguisherMethod: { + '#withType':: d.fn(help='`type` is the type of flow distinguisher method The supported types are "ByUser" and "ByNamespace". Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { distinguisherMethod+: { type: type } } }, + }, + '#priorityLevelConfiguration':: d.obj(help='PriorityLevelConfigurationReference contains information that points to the "request-priority" being used.'), + priorityLevelConfiguration: { + '#withName':: d.fn(help='`name` is the name of the priority level configuration being referenced Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { priorityLevelConfiguration+: { name: name } } }, + }, + '#withMatchingPrecedence':: d.fn(help='`matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen FlowSchema is among those with the numerically lowest (which we take to be logically highest) MatchingPrecedence. Each MatchingPrecedence value must be ranged in [1,10000]. Note that if the precedence is not specified, it will be set to 1000 as default.', args=[d.arg(name='matchingPrecedence', type=d.T.integer)]), + withMatchingPrecedence(matchingPrecedence): { spec+: { matchingPrecedence: matchingPrecedence } }, + '#withRules':: d.fn(help='`rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if at least one member of rules matches the request. if it is an empty slice, there will be no requests matching the FlowSchema.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { spec+: { rules: if std.isArray(v=rules) then rules else [rules] } }, + '#withRulesMixin':: d.fn(help='`rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if at least one member of rules matches the request. if it is an empty slice, there will be no requests matching the FlowSchema.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { spec+: { rules+: if std.isArray(v=rules) then rules else [rules] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaCondition.libsonnet new file mode 100644 index 00000000000..9c20d48508e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowSchemaCondition', url='', help='FlowSchemaCondition describes conditions for a FlowSchema.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='`message` is a human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="`reason` is a unique, one-word, CamelCase reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='`type` is the type of the condition. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaSpec.libsonnet new file mode 100644 index 00000000000..3c34b4a719b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaSpec.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowSchemaSpec', url='', help="FlowSchemaSpec describes how the FlowSchema's specification looks like."), + '#distinguisherMethod':: d.obj(help='FlowDistinguisherMethod specifies the method of a flow distinguisher.'), + distinguisherMethod: { + '#withType':: d.fn(help='`type` is the type of flow distinguisher method The supported types are "ByUser" and "ByNamespace". Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { distinguisherMethod+: { type: type } }, + }, + '#priorityLevelConfiguration':: d.obj(help='PriorityLevelConfigurationReference contains information that points to the "request-priority" being used.'), + priorityLevelConfiguration: { + '#withName':: d.fn(help='`name` is the name of the priority level configuration being referenced Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { priorityLevelConfiguration+: { name: name } }, + }, + '#withMatchingPrecedence':: d.fn(help='`matchingPrecedence` is used to choose among the FlowSchemas that match a given request. The chosen FlowSchema is among those with the numerically lowest (which we take to be logically highest) MatchingPrecedence. Each MatchingPrecedence value must be ranged in [1,10000]. Note that if the precedence is not specified, it will be set to 1000 as default.', args=[d.arg(name='matchingPrecedence', type=d.T.integer)]), + withMatchingPrecedence(matchingPrecedence): { matchingPrecedence: matchingPrecedence }, + '#withRules':: d.fn(help='`rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if at least one member of rules matches the request. if it is an empty slice, there will be no requests matching the FlowSchema.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='`rules` describes which requests will match this flow schema. This FlowSchema matches a request if and only if at least one member of rules matches the request. if it is an empty slice, there will be no requests matching the FlowSchema.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaStatus.libsonnet new file mode 100644 index 00000000000..2b43aa5e731 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/flowSchemaStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='flowSchemaStatus', url='', help='FlowSchemaStatus represents the current state of a FlowSchema.'), + '#withConditions':: d.fn(help='`conditions` is a list of the current states of FlowSchema.', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='`conditions` is a list of the current states of FlowSchema.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/groupSubject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/groupSubject.libsonnet new file mode 100644 index 00000000000..ab259ab5f2c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/groupSubject.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='groupSubject', url='', help='GroupSubject holds detailed information for group-kind subject.'), + '#withName':: d.fn(help='name is the user group that matches, or "*" to match all user groups. See https://github.com/kubernetes/apiserver/blob/master/pkg/authentication/user/user.go for some well-known group names. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitResponse.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitResponse.libsonnet new file mode 100644 index 00000000000..8be3d15d04f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitResponse.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='limitResponse', url='', help='LimitResponse defines how to handle requests that can not be executed right now.'), + '#queuing':: d.obj(help='QueuingConfiguration holds the configuration parameters for queuing'), + queuing: { + '#withHandSize':: d.fn(help="`handSize` is a small positive number that configures the shuffle sharding of requests into queues. When enqueuing a request at this priority level the request's flow identifier (a string pair) is hashed and the hash value is used to shuffle the list of queues and deal a hand of the size specified here. The request is put into one of the shortest queues in that hand. `handSize` must be no larger than `queues`, and should be significantly smaller (so that a few heavy flows do not saturate most of the queues). See the user-facing documentation for more extensive guidance on setting this field. This field has a default value of 8.", args=[d.arg(name='handSize', type=d.T.integer)]), + withHandSize(handSize): { queuing+: { handSize: handSize } }, + '#withQueueLengthLimit':: d.fn(help='`queueLengthLimit` is the maximum number of requests allowed to be waiting in a given queue of this priority level at a time; excess requests are rejected. This value must be positive. If not specified, it will be defaulted to 50.', args=[d.arg(name='queueLengthLimit', type=d.T.integer)]), + withQueueLengthLimit(queueLengthLimit): { queuing+: { queueLengthLimit: queueLengthLimit } }, + '#withQueues':: d.fn(help='`queues` is the number of queues for this priority level. The queues exist independently at each apiserver. The value must be positive. Setting it to 1 effectively precludes shufflesharding and thus makes the distinguisher method of associated flow schemas irrelevant. This field has a default value of 64.', args=[d.arg(name='queues', type=d.T.integer)]), + withQueues(queues): { queuing+: { queues: queues } }, + }, + '#withType':: d.fn(help='`type` is "Queue" or "Reject". "Queue" means that requests that can not be executed upon arrival are held in a queue until they can be executed or a queuing limit is reached. "Reject" means that requests that can not be executed upon arrival are rejected. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitedPriorityLevelConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitedPriorityLevelConfiguration.libsonnet new file mode 100644 index 00000000000..6b46a2ce3b5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/limitedPriorityLevelConfiguration.libsonnet @@ -0,0 +1,22 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='limitedPriorityLevelConfiguration', url='', help='LimitedPriorityLevelConfiguration specifies how to handle requests that are subject to limits. It addresses two issues:\n * How are requests for this priority level limited?\n * What should be done with requests that exceed the limit?'), + '#limitResponse':: d.obj(help='LimitResponse defines how to handle requests that can not be executed right now.'), + limitResponse: { + '#queuing':: d.obj(help='QueuingConfiguration holds the configuration parameters for queuing'), + queuing: { + '#withHandSize':: d.fn(help="`handSize` is a small positive number that configures the shuffle sharding of requests into queues. When enqueuing a request at this priority level the request's flow identifier (a string pair) is hashed and the hash value is used to shuffle the list of queues and deal a hand of the size specified here. The request is put into one of the shortest queues in that hand. `handSize` must be no larger than `queues`, and should be significantly smaller (so that a few heavy flows do not saturate most of the queues). See the user-facing documentation for more extensive guidance on setting this field. This field has a default value of 8.", args=[d.arg(name='handSize', type=d.T.integer)]), + withHandSize(handSize): { limitResponse+: { queuing+: { handSize: handSize } } }, + '#withQueueLengthLimit':: d.fn(help='`queueLengthLimit` is the maximum number of requests allowed to be waiting in a given queue of this priority level at a time; excess requests are rejected. This value must be positive. If not specified, it will be defaulted to 50.', args=[d.arg(name='queueLengthLimit', type=d.T.integer)]), + withQueueLengthLimit(queueLengthLimit): { limitResponse+: { queuing+: { queueLengthLimit: queueLengthLimit } } }, + '#withQueues':: d.fn(help='`queues` is the number of queues for this priority level. The queues exist independently at each apiserver. The value must be positive. Setting it to 1 effectively precludes shufflesharding and thus makes the distinguisher method of associated flow schemas irrelevant. This field has a default value of 64.', args=[d.arg(name='queues', type=d.T.integer)]), + withQueues(queues): { limitResponse+: { queuing+: { queues: queues } } }, + }, + '#withType':: d.fn(help='`type` is "Queue" or "Reject". "Queue" means that requests that can not be executed upon arrival are held in a queue until they can be executed or a queuing limit is reached. "Reject" means that requests that can not be executed upon arrival are rejected. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { limitResponse+: { type: type } }, + }, + '#withAssuredConcurrencyShares':: d.fn(help="`assuredConcurrencyShares` (ACS) configures the execution limit, which is a limit on the number of requests of this priority level that may be exeucting at a given time. ACS must be a positive number. The server's concurrency limit (SCL) is divided among the concurrency-controlled priority levels in proportion to their assured concurrency shares. This produces the assured concurrency value (ACV) --- the number of requests that may be executing at a time --- for each such priority level:\n\n ACV(l) = ceil( SCL * ACS(l) / ( sum[priority levels k] ACS(k) ) )\n\nbigger numbers of ACS mean more reserved concurrent requests (at the expense of every other PL). This field has a default value of 30.", args=[d.arg(name='assuredConcurrencyShares', type=d.T.integer)]), + withAssuredConcurrencyShares(assuredConcurrencyShares): { assuredConcurrencyShares: assuredConcurrencyShares }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..5116477b91f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/main.libsonnet @@ -0,0 +1,24 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + flowDistinguisherMethod: (import 'flowDistinguisherMethod.libsonnet'), + flowSchema: (import 'flowSchema.libsonnet'), + flowSchemaCondition: (import 'flowSchemaCondition.libsonnet'), + flowSchemaSpec: (import 'flowSchemaSpec.libsonnet'), + flowSchemaStatus: (import 'flowSchemaStatus.libsonnet'), + groupSubject: (import 'groupSubject.libsonnet'), + limitResponse: (import 'limitResponse.libsonnet'), + limitedPriorityLevelConfiguration: (import 'limitedPriorityLevelConfiguration.libsonnet'), + nonResourcePolicyRule: (import 'nonResourcePolicyRule.libsonnet'), + policyRulesWithSubjects: (import 'policyRulesWithSubjects.libsonnet'), + priorityLevelConfiguration: (import 'priorityLevelConfiguration.libsonnet'), + priorityLevelConfigurationCondition: (import 'priorityLevelConfigurationCondition.libsonnet'), + priorityLevelConfigurationReference: (import 'priorityLevelConfigurationReference.libsonnet'), + priorityLevelConfigurationSpec: (import 'priorityLevelConfigurationSpec.libsonnet'), + priorityLevelConfigurationStatus: (import 'priorityLevelConfigurationStatus.libsonnet'), + queuingConfiguration: (import 'queuingConfiguration.libsonnet'), + resourcePolicyRule: (import 'resourcePolicyRule.libsonnet'), + serviceAccountSubject: (import 'serviceAccountSubject.libsonnet'), + subject: (import 'subject.libsonnet'), + userSubject: (import 'userSubject.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/nonResourcePolicyRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/nonResourcePolicyRule.libsonnet new file mode 100644 index 00000000000..23a88ad4e64 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/nonResourcePolicyRule.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='nonResourcePolicyRule', url='', help='NonResourcePolicyRule is a predicate that matches non-resource requests according to their verb and the target non-resource URL. A NonResourcePolicyRule matches a request if and only if both (a) at least one member of verbs matches the request and (b) at least one member of nonResourceURLs matches the request.'), + '#withNonResourceURLs':: d.fn(help='`nonResourceURLs` is a set of url prefixes that a user should have access to and may not be empty. For example:\n - "/healthz" is legal\n - "/hea*" is illegal\n - "/hea" is legal but matches nothing\n - "/hea/*" also matches nothing\n - "/healthz/*" matches all per-component health checks.\n"*" matches all non-resource urls. if it is present, it must be the only entry. Required.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='`nonResourceURLs` is a set of url prefixes that a user should have access to and may not be empty. For example:\n - "/healthz" is legal\n - "/hea*" is illegal\n - "/hea" is legal but matches nothing\n - "/hea/*" also matches nothing\n - "/healthz/*" matches all per-component health checks.\n"*" matches all non-resource urls. if it is present, it must be the only entry. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withVerbs':: d.fn(help='`verbs` is a list of matching verbs and may not be empty. "*" matches all verbs. If it is present, it must be the only entry. Required.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='`verbs` is a list of matching verbs and may not be empty. "*" matches all verbs. If it is present, it must be the only entry. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/policyRulesWithSubjects.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/policyRulesWithSubjects.libsonnet new file mode 100644 index 00000000000..87a2ece5ce1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/policyRulesWithSubjects.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='policyRulesWithSubjects', url='', help='PolicyRulesWithSubjects prescribes a test that applies to a request to an apiserver. The test considers the subject making the request, the verb being requested, and the resource to be acted upon. This PolicyRulesWithSubjects matches a request if and only if both (a) at least one member of subjects matches the request and (b) at least one member of resourceRules or nonResourceRules matches the request.'), + '#withNonResourceRules':: d.fn(help='`nonResourceRules` is a list of NonResourcePolicyRules that identify matching requests according to their verb and the target non-resource URL.', args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRules(nonResourceRules): { nonResourceRules: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withNonResourceRulesMixin':: d.fn(help='`nonResourceRules` is a list of NonResourcePolicyRules that identify matching requests according to their verb and the target non-resource URL.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceRules', type=d.T.array)]), + withNonResourceRulesMixin(nonResourceRules): { nonResourceRules+: if std.isArray(v=nonResourceRules) then nonResourceRules else [nonResourceRules] }, + '#withResourceRules':: d.fn(help='`resourceRules` is a slice of ResourcePolicyRules that identify matching requests according to their verb and the target resource. At least one of `resourceRules` and `nonResourceRules` has to be non-empty.', args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRules(resourceRules): { resourceRules: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#withResourceRulesMixin':: d.fn(help='`resourceRules` is a slice of ResourcePolicyRules that identify matching requests according to their verb and the target resource. At least one of `resourceRules` and `nonResourceRules` has to be non-empty.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceRules', type=d.T.array)]), + withResourceRulesMixin(resourceRules): { resourceRules+: if std.isArray(v=resourceRules) then resourceRules else [resourceRules] }, + '#withSubjects':: d.fn(help='subjects is the list of normal user, serviceaccount, or group that this rule cares about. There must be at least one member in this slice. A slice that includes both the system:authenticated and system:unauthenticated user groups matches every request. Required.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='subjects is the list of normal user, serviceaccount, or group that this rule cares about. There must be at least one member in this slice. A slice that includes both the system:authenticated and system:unauthenticated user groups matches every request. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfiguration.libsonnet new file mode 100644 index 00000000000..7e101e43531 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfiguration.libsonnet @@ -0,0 +1,80 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityLevelConfiguration', url='', help='PriorityLevelConfiguration represents the configuration of a priority level.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Prioritylevelconfiguration', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'flowcontrol.apiserver.k8s.io/v1alpha1', + kind: 'PriorityLevelConfiguration', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PriorityLevelConfigurationSpec specifies the configuration of a priority level.'), + spec: { + '#limited':: d.obj(help='LimitedPriorityLevelConfiguration specifies how to handle requests that are subject to limits. It addresses two issues:\n * How are requests for this priority level limited?\n * What should be done with requests that exceed the limit?'), + limited: { + '#limitResponse':: d.obj(help='LimitResponse defines how to handle requests that can not be executed right now.'), + limitResponse: { + '#queuing':: d.obj(help='QueuingConfiguration holds the configuration parameters for queuing'), + queuing: { + '#withHandSize':: d.fn(help="`handSize` is a small positive number that configures the shuffle sharding of requests into queues. When enqueuing a request at this priority level the request's flow identifier (a string pair) is hashed and the hash value is used to shuffle the list of queues and deal a hand of the size specified here. The request is put into one of the shortest queues in that hand. `handSize` must be no larger than `queues`, and should be significantly smaller (so that a few heavy flows do not saturate most of the queues). See the user-facing documentation for more extensive guidance on setting this field. This field has a default value of 8.", args=[d.arg(name='handSize', type=d.T.integer)]), + withHandSize(handSize): { spec+: { limited+: { limitResponse+: { queuing+: { handSize: handSize } } } } }, + '#withQueueLengthLimit':: d.fn(help='`queueLengthLimit` is the maximum number of requests allowed to be waiting in a given queue of this priority level at a time; excess requests are rejected. This value must be positive. If not specified, it will be defaulted to 50.', args=[d.arg(name='queueLengthLimit', type=d.T.integer)]), + withQueueLengthLimit(queueLengthLimit): { spec+: { limited+: { limitResponse+: { queuing+: { queueLengthLimit: queueLengthLimit } } } } }, + '#withQueues':: d.fn(help='`queues` is the number of queues for this priority level. The queues exist independently at each apiserver. The value must be positive. Setting it to 1 effectively precludes shufflesharding and thus makes the distinguisher method of associated flow schemas irrelevant. This field has a default value of 64.', args=[d.arg(name='queues', type=d.T.integer)]), + withQueues(queues): { spec+: { limited+: { limitResponse+: { queuing+: { queues: queues } } } } }, + }, + '#withType':: d.fn(help='`type` is "Queue" or "Reject". "Queue" means that requests that can not be executed upon arrival are held in a queue until they can be executed or a queuing limit is reached. "Reject" means that requests that can not be executed upon arrival are rejected. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { limited+: { limitResponse+: { type: type } } } }, + }, + '#withAssuredConcurrencyShares':: d.fn(help="`assuredConcurrencyShares` (ACS) configures the execution limit, which is a limit on the number of requests of this priority level that may be exeucting at a given time. ACS must be a positive number. The server's concurrency limit (SCL) is divided among the concurrency-controlled priority levels in proportion to their assured concurrency shares. This produces the assured concurrency value (ACV) --- the number of requests that may be executing at a time --- for each such priority level:\n\n ACV(l) = ceil( SCL * ACS(l) / ( sum[priority levels k] ACS(k) ) )\n\nbigger numbers of ACS mean more reserved concurrent requests (at the expense of every other PL). This field has a default value of 30.", args=[d.arg(name='assuredConcurrencyShares', type=d.T.integer)]), + withAssuredConcurrencyShares(assuredConcurrencyShares): { spec+: { limited+: { assuredConcurrencyShares: assuredConcurrencyShares } } }, + }, + '#withType':: d.fn(help="`type` indicates whether this priority level is subject to limitation on request execution. A value of `'Exempt'` means that requests of this priority level are not subject to a limit (and thus are never queued) and do not detract from the capacity made available to other priority levels. A value of `'Limited'` means that (a) requests of this priority level _are_ subject to limits and (b) some of the server's limited capacity is made available exclusively to this priority level. Required.", args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { type: type } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationCondition.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationCondition.libsonnet new file mode 100644 index 00000000000..db0280bc1ba --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationCondition.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityLevelConfigurationCondition', url='', help='PriorityLevelConfigurationCondition defines the condition of priority level.'), + '#withLastTransitionTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='lastTransitionTime', type=d.T.string)]), + withLastTransitionTime(lastTransitionTime): { lastTransitionTime: lastTransitionTime }, + '#withMessage':: d.fn(help='`message` is a human-readable message indicating details about last transition.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withReason':: d.fn(help="`reason` is a unique, one-word, CamelCase reason for the condition's last transition.", args=[d.arg(name='reason', type=d.T.string)]), + withReason(reason): { reason: reason }, + '#withType':: d.fn(help='`type` is the type of the condition. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationReference.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationReference.libsonnet new file mode 100644 index 00000000000..535f0a18e72 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationReference.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityLevelConfigurationReference', url='', help='PriorityLevelConfigurationReference contains information that points to the "request-priority" being used.'), + '#withName':: d.fn(help='`name` is the name of the priority level configuration being referenced Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationSpec.libsonnet new file mode 100644 index 00000000000..6771cd0e7e4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationSpec.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityLevelConfigurationSpec', url='', help='PriorityLevelConfigurationSpec specifies the configuration of a priority level.'), + '#limited':: d.obj(help='LimitedPriorityLevelConfiguration specifies how to handle requests that are subject to limits. It addresses two issues:\n * How are requests for this priority level limited?\n * What should be done with requests that exceed the limit?'), + limited: { + '#limitResponse':: d.obj(help='LimitResponse defines how to handle requests that can not be executed right now.'), + limitResponse: { + '#queuing':: d.obj(help='QueuingConfiguration holds the configuration parameters for queuing'), + queuing: { + '#withHandSize':: d.fn(help="`handSize` is a small positive number that configures the shuffle sharding of requests into queues. When enqueuing a request at this priority level the request's flow identifier (a string pair) is hashed and the hash value is used to shuffle the list of queues and deal a hand of the size specified here. The request is put into one of the shortest queues in that hand. `handSize` must be no larger than `queues`, and should be significantly smaller (so that a few heavy flows do not saturate most of the queues). See the user-facing documentation for more extensive guidance on setting this field. This field has a default value of 8.", args=[d.arg(name='handSize', type=d.T.integer)]), + withHandSize(handSize): { limited+: { limitResponse+: { queuing+: { handSize: handSize } } } }, + '#withQueueLengthLimit':: d.fn(help='`queueLengthLimit` is the maximum number of requests allowed to be waiting in a given queue of this priority level at a time; excess requests are rejected. This value must be positive. If not specified, it will be defaulted to 50.', args=[d.arg(name='queueLengthLimit', type=d.T.integer)]), + withQueueLengthLimit(queueLengthLimit): { limited+: { limitResponse+: { queuing+: { queueLengthLimit: queueLengthLimit } } } }, + '#withQueues':: d.fn(help='`queues` is the number of queues for this priority level. The queues exist independently at each apiserver. The value must be positive. Setting it to 1 effectively precludes shufflesharding and thus makes the distinguisher method of associated flow schemas irrelevant. This field has a default value of 64.', args=[d.arg(name='queues', type=d.T.integer)]), + withQueues(queues): { limited+: { limitResponse+: { queuing+: { queues: queues } } } }, + }, + '#withType':: d.fn(help='`type` is "Queue" or "Reject". "Queue" means that requests that can not be executed upon arrival are held in a queue until they can be executed or a queuing limit is reached. "Reject" means that requests that can not be executed upon arrival are rejected. Required.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { limited+: { limitResponse+: { type: type } } }, + }, + '#withAssuredConcurrencyShares':: d.fn(help="`assuredConcurrencyShares` (ACS) configures the execution limit, which is a limit on the number of requests of this priority level that may be exeucting at a given time. ACS must be a positive number. The server's concurrency limit (SCL) is divided among the concurrency-controlled priority levels in proportion to their assured concurrency shares. This produces the assured concurrency value (ACV) --- the number of requests that may be executing at a time --- for each such priority level:\n\n ACV(l) = ceil( SCL * ACS(l) / ( sum[priority levels k] ACS(k) ) )\n\nbigger numbers of ACS mean more reserved concurrent requests (at the expense of every other PL). This field has a default value of 30.", args=[d.arg(name='assuredConcurrencyShares', type=d.T.integer)]), + withAssuredConcurrencyShares(assuredConcurrencyShares): { limited+: { assuredConcurrencyShares: assuredConcurrencyShares } }, + }, + '#withType':: d.fn(help="`type` indicates whether this priority level is subject to limitation on request execution. A value of `'Exempt'` means that requests of this priority level are not subject to a limit (and thus are never queued) and do not detract from the capacity made available to other priority levels. A value of `'Limited'` means that (a) requests of this priority level _are_ subject to limits and (b) some of the server's limited capacity is made available exclusively to this priority level. Required.", args=[d.arg(name='type', type=d.T.string)]), + withType(type): { type: type }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationStatus.libsonnet new file mode 100644 index 00000000000..b3c97ed257b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/priorityLevelConfigurationStatus.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityLevelConfigurationStatus', url='', help='PriorityLevelConfigurationStatus represents the current state of a "request-priority".'), + '#withConditions':: d.fn(help='`conditions` is the current state of "request-priority".', args=[d.arg(name='conditions', type=d.T.array)]), + withConditions(conditions): { conditions: if std.isArray(v=conditions) then conditions else [conditions] }, + '#withConditionsMixin':: d.fn(help='`conditions` is the current state of "request-priority".\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='conditions', type=d.T.array)]), + withConditionsMixin(conditions): { conditions+: if std.isArray(v=conditions) then conditions else [conditions] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/queuingConfiguration.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/queuingConfiguration.libsonnet new file mode 100644 index 00000000000..4f72776a215 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/queuingConfiguration.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='queuingConfiguration', url='', help='QueuingConfiguration holds the configuration parameters for queuing'), + '#withHandSize':: d.fn(help="`handSize` is a small positive number that configures the shuffle sharding of requests into queues. When enqueuing a request at this priority level the request's flow identifier (a string pair) is hashed and the hash value is used to shuffle the list of queues and deal a hand of the size specified here. The request is put into one of the shortest queues in that hand. `handSize` must be no larger than `queues`, and should be significantly smaller (so that a few heavy flows do not saturate most of the queues). See the user-facing documentation for more extensive guidance on setting this field. This field has a default value of 8.", args=[d.arg(name='handSize', type=d.T.integer)]), + withHandSize(handSize): { handSize: handSize }, + '#withQueueLengthLimit':: d.fn(help='`queueLengthLimit` is the maximum number of requests allowed to be waiting in a given queue of this priority level at a time; excess requests are rejected. This value must be positive. If not specified, it will be defaulted to 50.', args=[d.arg(name='queueLengthLimit', type=d.T.integer)]), + withQueueLengthLimit(queueLengthLimit): { queueLengthLimit: queueLengthLimit }, + '#withQueues':: d.fn(help='`queues` is the number of queues for this priority level. The queues exist independently at each apiserver. The value must be positive. Setting it to 1 effectively precludes shufflesharding and thus makes the distinguisher method of associated flow schemas irrelevant. This field has a default value of 64.', args=[d.arg(name='queues', type=d.T.integer)]), + withQueues(queues): { queues: queues }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/resourcePolicyRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/resourcePolicyRule.libsonnet new file mode 100644 index 00000000000..df304dd21e7 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/resourcePolicyRule.libsonnet @@ -0,0 +1,24 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='resourcePolicyRule', url='', help="ResourcePolicyRule is a predicate that matches some resource requests, testing the request's verb and the target resource. A ResourcePolicyRule matches a resource request if and only if: (a) at least one member of verbs matches the request, (b) at least one member of apiGroups matches the request, (c) at least one member of resources matches the request, and (d) least one member of namespaces matches the request."), + '#withApiGroups':: d.fn(help='`apiGroups` is a list of matching API groups and may not be empty. "*" matches all API groups and, if present, must be the only entry. Required.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='`apiGroups` is a list of matching API groups and may not be empty. "*" matches all API groups and, if present, must be the only entry. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withClusterScope':: d.fn(help='`clusterScope` indicates whether to match requests that do not specify a namespace (which happens either because the resource is not namespaced or the request targets all namespaces). If this field is omitted or false then the `namespaces` field must contain a non-empty list.', args=[d.arg(name='clusterScope', type=d.T.boolean)]), + withClusterScope(clusterScope): { clusterScope: clusterScope }, + '#withNamespaces':: d.fn(help='`namespaces` is a list of target namespaces that restricts matches. A request that specifies a target namespace matches only if either (a) this list contains that target namespace or (b) this list contains "*". Note that "*" matches any specified namespace but does not match a request that _does not specify_ a namespace (see the `clusterScope` field for that). This list may be empty, but only if `clusterScope` is true.', args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help='`namespaces` is a list of target namespaces that restricts matches. A request that specifies a target namespace matches only if either (a) this list contains that target namespace or (b) this list contains "*". Note that "*" matches any specified namespace but does not match a request that _does not specify_ a namespace (see the `clusterScope` field for that). This list may be empty, but only if `clusterScope` is true.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withResources':: d.fn(help='`resources` is a list of matching resources (i.e., lowercase and plural) with, if desired, subresource. For example, [ "services", "nodes/status" ]. This list may not be empty. "*" matches all resources and, if present, must be the only entry. Required.', args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help='`resources` is a list of matching resources (i.e., lowercase and plural) with, if desired, subresource. For example, [ "services", "nodes/status" ]. This list may not be empty. "*" matches all resources and, if present, must be the only entry. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='`verbs` is a list of matching verbs and may not be empty. "*" matches all verbs and, if present, must be the only entry. Required.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='`verbs` is a list of matching verbs and may not be empty. "*" matches all verbs and, if present, must be the only entry. Required.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/serviceAccountSubject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/serviceAccountSubject.libsonnet new file mode 100644 index 00000000000..836b2408317 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/serviceAccountSubject.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceAccountSubject', url='', help='ServiceAccountSubject holds detailed information for service-account-kind subject.'), + '#withName':: d.fn(help='`name` is the name of matching ServiceAccount objects, or "*" to match regardless of name. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of matching ServiceAccount objects. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/subject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/subject.libsonnet new file mode 100644 index 00000000000..0093b04c852 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/subject.libsonnet @@ -0,0 +1,25 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subject', url='', help='Subject matches the originator of a request, as identified by the request authentication system. There are three ways of matching an originator; by user, group, or service account.'), + '#group':: d.obj(help='GroupSubject holds detailed information for group-kind subject.'), + group: { + '#withName':: d.fn(help='name is the user group that matches, or "*" to match all user groups. See https://github.com/kubernetes/apiserver/blob/master/pkg/authentication/user/user.go for some well-known group names. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { group+: { name: name } }, + }, + '#serviceAccount':: d.obj(help='ServiceAccountSubject holds detailed information for service-account-kind subject.'), + serviceAccount: { + '#withName':: d.fn(help='`name` is the name of matching ServiceAccount objects, or "*" to match regardless of name. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { serviceAccount+: { name: name } }, + '#withNamespace':: d.fn(help='`namespace` is the namespace of matching ServiceAccount objects. Required.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { serviceAccount+: { namespace: namespace } }, + }, + '#user':: d.obj(help='UserSubject holds detailed information for user-kind subject.'), + user: { + '#withName':: d.fn(help='`name` is the username that matches, or "*" to match all usernames. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { user+: { name: name } }, + }, + '#withKind':: d.fn(help='Required', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/userSubject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/userSubject.libsonnet new file mode 100644 index 00000000000..e4f076ace16 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/flowcontrol/v1alpha1/userSubject.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='userSubject', url='', help='UserSubject holds detailed information for user-kind subject.'), + '#withName':: d.fn(help='`name` is the username that matches, or "*" to match all usernames. Required.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/main.libsonnet new file mode 100644 index 00000000000..48137d12473 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networking', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressPath.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressPath.libsonnet new file mode 100644 index 00000000000..a65072d68d2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressPath.libsonnet @@ -0,0 +1,34 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressPath', url='', help='HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.'), + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { backend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { backend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { resource+: { name: name } } }, + }, + '#service':: d.obj(help='IngressServiceBackend references a Kubernetes Service as a Backend.'), + service: { + '#port':: d.obj(help='ServiceBackendPort is the service port being referenced.'), + port: { + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { service+: { port+: { name: name } } } }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { backend+: { service+: { port+: { number: number } } } }, + }, + '#withName':: d.fn(help='Name is the referenced service. The service must exist in the same namespace as the Ingress object.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { service+: { name: name } } }, + }, + }, + '#withPath':: d.fn(help="Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional 'path' part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPathType':: d.fn(help="PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types.", args=[d.arg(name='pathType', type=d.T.string)]), + withPathType(pathType): { pathType: pathType }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressRuleValue.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressRuleValue.libsonnet new file mode 100644 index 00000000000..31169937a55 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/httpIngressRuleValue.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressRuleValue', url='', help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { paths: if std.isArray(v=paths) then paths else [paths] }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { paths+: if std.isArray(v=paths) then paths else [paths] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingress.libsonnet new file mode 100644 index 00000000000..db772c77ad6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingress.libsonnet @@ -0,0 +1,93 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingress', url='', help='Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Ingress', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'networking.k8s.io/v1', + kind: 'Ingress', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='IngressSpec describes the Ingress the user wishes to exist.'), + spec: { + '#defaultBackend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + defaultBackend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { defaultBackend+: { resource+: { apiGroup: apiGroup } } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { defaultBackend+: { resource+: { kind: kind } } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { defaultBackend+: { resource+: { name: name } } } }, + }, + '#service':: d.obj(help='IngressServiceBackend references a Kubernetes Service as a Backend.'), + service: { + '#port':: d.obj(help='ServiceBackendPort is the service port being referenced.'), + port: { + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { defaultBackend+: { service+: { port+: { name: name } } } } }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { spec+: { defaultBackend+: { service+: { port+: { number: number } } } } }, + }, + '#withName':: d.fn(help='Name is the referenced service. The service must exist in the same namespace as the Ingress object.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { defaultBackend+: { service+: { name: name } } } }, + }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { spec+: { ingressClassName: ingressClassName } }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { spec+: { rules: if std.isArray(v=rules) then rules else [rules] } }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { spec+: { rules+: if std.isArray(v=rules) then rules else [rules] } }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { spec+: { tls: if std.isArray(v=tls) then tls else [tls] } }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { spec+: { tls+: if std.isArray(v=tls) then tls else [tls] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressBackend.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressBackend.libsonnet new file mode 100644 index 00000000000..e480a979691 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressBackend.libsonnet @@ -0,0 +1,27 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressBackend', url='', help='IngressBackend describes all endpoints for a given service and port.'), + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { resource+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { resource+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#service':: d.obj(help='IngressServiceBackend references a Kubernetes Service as a Backend.'), + service: { + '#port':: d.obj(help='ServiceBackendPort is the service port being referenced.'), + port: { + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { port+: { name: name } } }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { service+: { port+: { number: number } } }, + }, + '#withName':: d.fn(help='Name is the referenced service. The service must exist in the same namespace as the Ingress object.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { service+: { name: name } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClass.libsonnet new file mode 100644 index 00000000000..0fe56a04851 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClass.libsonnet @@ -0,0 +1,70 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressClass', url='', help='IngressClass represents the class of the Ingress, referenced by the Ingress Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be used to indicate that an IngressClass should be considered default. When a single IngressClass resource has this annotation set to true, new Ingress resources without a class specified will be assigned this default class.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Ingressclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'networking.k8s.io/v1', + kind: 'IngressClass', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='IngressClassSpec provides information about the class of an Ingress.'), + spec: { + '#parameters':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + parameters: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { parameters+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { parameters+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { parameters+: { name: name } } }, + }, + '#withController':: d.fn(help='Controller refers to the name of the controller that should handle this class. This allows for different "flavors" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. "acme.io/ingress-controller". This field is immutable.', args=[d.arg(name='controller', type=d.T.string)]), + withController(controller): { spec+: { controller: controller } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClassSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClassSpec.libsonnet new file mode 100644 index 00000000000..73595cd1caa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressClassSpec.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressClassSpec', url='', help='IngressClassSpec provides information about the class of an Ingress.'), + '#parameters':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + parameters: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { parameters+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { parameters+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { parameters+: { name: name } }, + }, + '#withController':: d.fn(help='Controller refers to the name of the controller that should handle this class. This allows for different "flavors" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. "acme.io/ingress-controller". This field is immutable.', args=[d.arg(name='controller', type=d.T.string)]), + withController(controller): { controller: controller }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressRule.libsonnet new file mode 100644 index 00000000000..3a0a5dca509 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressRule.libsonnet @@ -0,0 +1,15 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressRule', url='', help='IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.'), + '#http':: d.obj(help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + http: { + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { http+: { paths: if std.isArray(v=paths) then paths else [paths] } }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { http+: { paths+: if std.isArray(v=paths) then paths else [paths] } }, + }, + '#withHost':: d.fn(help="Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the 'host' part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be 'precise' which is a domain name without the terminating dot of a network host (e.g. 'foo.bar.com') or 'wildcard', which is a domain name prefixed with a single wildcard label (e.g. '*.foo.com'). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == '*'). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressServiceBackend.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressServiceBackend.libsonnet new file mode 100644 index 00000000000..93ba846cf91 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressServiceBackend.libsonnet @@ -0,0 +1,15 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressServiceBackend', url='', help='IngressServiceBackend references a Kubernetes Service as a Backend.'), + '#port':: d.obj(help='ServiceBackendPort is the service port being referenced.'), + port: { + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { port+: { name: name } }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { port+: { number: number } }, + }, + '#withName':: d.fn(help='Name is the referenced service. The service must exist in the same namespace as the Ingress object.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressSpec.libsonnet new file mode 100644 index 00000000000..77acb03787d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressSpec.libsonnet @@ -0,0 +1,40 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressSpec', url='', help='IngressSpec describes the Ingress the user wishes to exist.'), + '#defaultBackend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + defaultBackend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { defaultBackend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { defaultBackend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { defaultBackend+: { resource+: { name: name } } }, + }, + '#service':: d.obj(help='IngressServiceBackend references a Kubernetes Service as a Backend.'), + service: { + '#port':: d.obj(help='ServiceBackendPort is the service port being referenced.'), + port: { + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { defaultBackend+: { service+: { port+: { name: name } } } }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { defaultBackend+: { service+: { port+: { number: number } } } }, + }, + '#withName':: d.fn(help='Name is the referenced service. The service must exist in the same namespace as the Ingress object.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { defaultBackend+: { service+: { name: name } } }, + }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { ingressClassName: ingressClassName }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { tls: if std.isArray(v=tls) then tls else [tls] }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { tls+: if std.isArray(v=tls) then tls else [tls] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressStatus.libsonnet new file mode 100644 index 00000000000..697b5b33c1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressStatus.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressStatus', url='', help='IngressStatus describe the current state of the Ingress.'), + '#loadBalancer':: d.obj(help='LoadBalancerStatus represents the status of a load-balancer.'), + loadBalancer: { + '#withIngress':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.', args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { loadBalancer+: { ingress: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withIngressMixin':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { loadBalancer+: { ingress+: if std.isArray(v=ingress) then ingress else [ingress] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressTLS.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressTLS.libsonnet new file mode 100644 index 00000000000..fbbe8558c31 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ingressTLS.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressTLS', url='', help='IngressTLS describes the transport layer security associated with an Ingress.'), + '#withHosts':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.', args=[d.arg(name='hosts', type=d.T.array)]), + withHosts(hosts): { hosts: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withHostsMixin':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hosts', type=d.T.array)]), + withHostsMixin(hosts): { hosts+: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withSecretName':: d.fn(help='SecretName is the name of the secret used to terminate TLS traffic on port 443. Field is left optional to allow TLS routing based on SNI hostname alone. If the SNI host in a listener conflicts with the "Host" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ipBlock.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ipBlock.libsonnet new file mode 100644 index 00000000000..b7523163cce --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/ipBlock.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ipBlock', url='', help="IPBlock describes a particular CIDR (Ex. '192.168.1.1/24','2001:db9::/64') that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule."), + '#withCidr':: d.fn(help='CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64"', args=[d.arg(name='cidr', type=d.T.string)]), + withCidr(cidr): { cidr: cidr }, + '#withExcept':: d.fn(help='Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64" Except values will be rejected if they are outside the CIDR range', args=[d.arg(name='except', type=d.T.array)]), + withExcept(except): { except: if std.isArray(v=except) then except else [except] }, + '#withExceptMixin':: d.fn(help='Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64" Except values will be rejected if they are outside the CIDR range\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='except', type=d.T.array)]), + withExceptMixin(except): { except+: if std.isArray(v=except) then except else [except] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/main.libsonnet new file mode 100644 index 00000000000..2c517b4c71a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/main.libsonnet @@ -0,0 +1,23 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + httpIngressPath: (import 'httpIngressPath.libsonnet'), + httpIngressRuleValue: (import 'httpIngressRuleValue.libsonnet'), + ingress: (import 'ingress.libsonnet'), + ingressBackend: (import 'ingressBackend.libsonnet'), + ingressClass: (import 'ingressClass.libsonnet'), + ingressClassSpec: (import 'ingressClassSpec.libsonnet'), + ingressRule: (import 'ingressRule.libsonnet'), + ingressServiceBackend: (import 'ingressServiceBackend.libsonnet'), + ingressSpec: (import 'ingressSpec.libsonnet'), + ingressStatus: (import 'ingressStatus.libsonnet'), + ingressTLS: (import 'ingressTLS.libsonnet'), + ipBlock: (import 'ipBlock.libsonnet'), + networkPolicy: (import 'networkPolicy.libsonnet'), + networkPolicyEgressRule: (import 'networkPolicyEgressRule.libsonnet'), + networkPolicyIngressRule: (import 'networkPolicyIngressRule.libsonnet'), + networkPolicyPeer: (import 'networkPolicyPeer.libsonnet'), + networkPolicyPort: (import 'networkPolicyPort.libsonnet'), + networkPolicySpec: (import 'networkPolicySpec.libsonnet'), + serviceBackendPort: (import 'serviceBackendPort.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicy.libsonnet new file mode 100644 index 00000000000..be73492a0d3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicy.libsonnet @@ -0,0 +1,82 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicy', url='', help='NetworkPolicy describes what network traffic is allowed for a set of Pods'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Networkpolicy', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'networking.k8s.io/v1', + kind: 'NetworkPolicy', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='NetworkPolicySpec provides the specification of a NetworkPolicy'), + spec: { + '#podSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + podSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { podSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { podSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { podSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { podSelector+: { matchLabels+: matchLabels } } }, + }, + '#withEgress':: d.fn(help='List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8', args=[d.arg(name='egress', type=d.T.array)]), + withEgress(egress): { spec+: { egress: if std.isArray(v=egress) then egress else [egress] } }, + '#withEgressMixin':: d.fn(help='List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='egress', type=d.T.array)]), + withEgressMixin(egress): { spec+: { egress+: if std.isArray(v=egress) then egress else [egress] } }, + '#withIngress':: d.fn(help="List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default)", args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { spec+: { ingress: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withIngressMixin':: d.fn(help="List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default)\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { spec+: { ingress+: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withPolicyTypes':: d.fn(help='List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8', args=[d.arg(name='policyTypes', type=d.T.array)]), + withPolicyTypes(policyTypes): { spec+: { policyTypes: if std.isArray(v=policyTypes) then policyTypes else [policyTypes] } }, + '#withPolicyTypesMixin':: d.fn(help='List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policyTypes', type=d.T.array)]), + withPolicyTypesMixin(policyTypes): { spec+: { policyTypes+: if std.isArray(v=policyTypes) then policyTypes else [policyTypes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyEgressRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyEgressRule.libsonnet new file mode 100644 index 00000000000..eb4f7f9bbd2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyEgressRule.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicyEgressRule', url='', help="NetworkPolicyEgressRule describes a particular set of traffic that is allowed out of pods matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and to. This type is beta-level in 1.8"), + '#withPorts':: d.fn(help='List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withTo':: d.fn(help='List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list.', args=[d.arg(name='to', type=d.T.array)]), + withTo(to): { to: if std.isArray(v=to) then to else [to] }, + '#withToMixin':: d.fn(help='List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='to', type=d.T.array)]), + withToMixin(to): { to+: if std.isArray(v=to) then to else [to] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyIngressRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyIngressRule.libsonnet new file mode 100644 index 00000000000..eb84a4ac05d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyIngressRule.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicyIngressRule', url='', help="NetworkPolicyIngressRule describes a particular set of traffic that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and from."), + '#withFrom':: d.fn(help='List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the from list.', args=[d.arg(name='from', type=d.T.array)]), + withFrom(from): { from: if std.isArray(v=from) then from else [from] }, + '#withFromMixin':: d.fn(help='List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the from list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='from', type=d.T.array)]), + withFromMixin(from): { from+: if std.isArray(v=from) then from else [from] }, + '#withPorts':: d.fn(help='List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list.', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPeer.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPeer.libsonnet new file mode 100644 index 00000000000..625d2a5f0f9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPeer.libsonnet @@ -0,0 +1,37 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicyPeer', url='', help='NetworkPolicyPeer describes a peer to allow traffic to/from. Only certain combinations of fields are allowed'), + '#ipBlock':: d.obj(help="IPBlock describes a particular CIDR (Ex. '192.168.1.1/24','2001:db9::/64') that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule."), + ipBlock: { + '#withCidr':: d.fn(help='CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64"', args=[d.arg(name='cidr', type=d.T.string)]), + withCidr(cidr): { ipBlock+: { cidr: cidr } }, + '#withExcept':: d.fn(help='Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64" Except values will be rejected if they are outside the CIDR range', args=[d.arg(name='except', type=d.T.array)]), + withExcept(except): { ipBlock+: { except: if std.isArray(v=except) then except else [except] } }, + '#withExceptMixin':: d.fn(help='Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" or "2001:db9::/64" Except values will be rejected if they are outside the CIDR range\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='except', type=d.T.array)]), + withExceptMixin(except): { ipBlock+: { except+: if std.isArray(v=except) then except else [except] } }, + }, + '#namespaceSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + namespaceSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#podSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + podSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podSelector+: { matchLabels+: matchLabels } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPort.libsonnet new file mode 100644 index 00000000000..e0905fc8188 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicyPort.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicyPort', url='', help='NetworkPolicyPort describes a port to allow traffic on'), + '#withPort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { port: port }, + '#withProtocol':: d.fn(help='The protocol (TCP, UDP, or SCTP) which traffic must match. If not specified, this field defaults to TCP.', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicySpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicySpec.libsonnet new file mode 100644 index 00000000000..c2e4c3af6f6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/networkPolicySpec.libsonnet @@ -0,0 +1,29 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='networkPolicySpec', url='', help='NetworkPolicySpec provides the specification of a NetworkPolicy'), + '#podSelector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + podSelector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podSelector+: { matchLabels+: matchLabels } }, + }, + '#withEgress':: d.fn(help='List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8', args=[d.arg(name='egress', type=d.T.array)]), + withEgress(egress): { egress: if std.isArray(v=egress) then egress else [egress] }, + '#withEgressMixin':: d.fn(help='List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='egress', type=d.T.array)]), + withEgressMixin(egress): { egress+: if std.isArray(v=egress) then egress else [egress] }, + '#withIngress':: d.fn(help="List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default)", args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { ingress: if std.isArray(v=ingress) then ingress else [ingress] }, + '#withIngressMixin':: d.fn(help="List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default)\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { ingress+: if std.isArray(v=ingress) then ingress else [ingress] }, + '#withPolicyTypes':: d.fn(help='List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8', args=[d.arg(name='policyTypes', type=d.T.array)]), + withPolicyTypes(policyTypes): { policyTypes: if std.isArray(v=policyTypes) then policyTypes else [policyTypes] }, + '#withPolicyTypesMixin':: d.fn(help='List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='policyTypes', type=d.T.array)]), + withPolicyTypesMixin(policyTypes): { policyTypes+: if std.isArray(v=policyTypes) then policyTypes else [policyTypes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/serviceBackendPort.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/serviceBackendPort.libsonnet new file mode 100644 index 00000000000..3688de8126b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1/serviceBackendPort.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='serviceBackendPort', url='', help='ServiceBackendPort is the service port being referenced.'), + '#withName':: d.fn(help='Name is the name of the port on the Service. This is a mutually exclusive setting with "Number".', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNumber':: d.fn(help='Number is the numerical port number (e.g. 80) on the Service. This is a mutually exclusive setting with "Name".', args=[d.arg(name='number', type=d.T.integer)]), + withNumber(number): { number: number }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressPath.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressPath.libsonnet new file mode 100644 index 00000000000..75e77b0c68d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressPath.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressPath', url='', help='HTTPIngressPath associates a path with a backend. Incoming urls matching the path are forwarded to the backend.'), + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { backend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { backend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { resource+: { name: name } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { backend+: { serviceName: serviceName } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { backend+: { servicePort: servicePort } }, + }, + '#withPath':: d.fn(help="Path is matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional 'path' part of a URL as defined by RFC 3986. Paths must begin with a '/'. When unspecified, all paths from incoming requests are matched.", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPathType':: d.fn(help="PathType determines the interpretation of the Path matching. PathType can be one of the following values: * Exact: Matches the URL path exactly. * Prefix: Matches based on a URL path prefix split by '/'. Matching is\n done on a path element by element basis. A path element refers is the\n list of labels in the path split by the '/' separator. A request is a\n match for path p if every p is an element-wise prefix of p of the\n request path. Note that if the last element of the path is a substring\n of the last element in request path, it is not a match (e.g. /foo/bar\n matches /foo/bar/baz, but does not match /foo/barbaz).\n* ImplementationSpecific: Interpretation of the Path matching is up to\n the IngressClass. Implementations can treat this as a separate PathType\n or treat it identically to Prefix or Exact path types.\nImplementations are required to support all path types. Defaults to ImplementationSpecific.", args=[d.arg(name='pathType', type=d.T.string)]), + withPathType(pathType): { pathType: pathType }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressRuleValue.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressRuleValue.libsonnet new file mode 100644 index 00000000000..31169937a55 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/httpIngressRuleValue.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='httpIngressRuleValue', url='', help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { paths: if std.isArray(v=paths) then paths else [paths] }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { paths+: if std.isArray(v=paths) then paths else [paths] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingress.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingress.libsonnet new file mode 100644 index 00000000000..5aeb1ccf620 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingress.libsonnet @@ -0,0 +1,85 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingress', url='', help='Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Ingress', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'networking.k8s.io/v1beta1', + kind: 'Ingress', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='IngressSpec describes the Ingress the user wishes to exist.'), + spec: { + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { backend+: { resource+: { apiGroup: apiGroup } } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { backend+: { resource+: { kind: kind } } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { backend+: { resource+: { name: name } } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { spec+: { backend+: { serviceName: serviceName } } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { spec+: { backend+: { servicePort: servicePort } } }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { spec+: { ingressClassName: ingressClassName } }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { spec+: { rules: if std.isArray(v=rules) then rules else [rules] } }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { spec+: { rules+: if std.isArray(v=rules) then rules else [rules] } }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { spec+: { tls: if std.isArray(v=tls) then tls else [tls] } }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { spec+: { tls+: if std.isArray(v=tls) then tls else [tls] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressBackend.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressBackend.libsonnet new file mode 100644 index 00000000000..1106f316187 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressBackend.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressBackend', url='', help='IngressBackend describes all endpoints for a given service and port.'), + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { resource+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { resource+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { resource+: { name: name } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { serviceName: serviceName }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { servicePort: servicePort }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClass.libsonnet new file mode 100644 index 00000000000..7c9eae7de70 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClass.libsonnet @@ -0,0 +1,70 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressClass', url='', help='IngressClass represents the class of the Ingress, referenced by the Ingress Spec. The `ingressclass.kubernetes.io/is-default-class` annotation can be used to indicate that an IngressClass should be considered default. When a single IngressClass resource has this annotation set to true, new Ingress resources without a class specified will be assigned this default class.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Ingressclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'networking.k8s.io/v1beta1', + kind: 'IngressClass', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='IngressClassSpec provides information about the class of an Ingress.'), + spec: { + '#parameters':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + parameters: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { parameters+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { parameters+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { parameters+: { name: name } } }, + }, + '#withController':: d.fn(help='Controller refers to the name of the controller that should handle this class. This allows for different "flavors" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. "acme.io/ingress-controller". This field is immutable.', args=[d.arg(name='controller', type=d.T.string)]), + withController(controller): { spec+: { controller: controller } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClassSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClassSpec.libsonnet new file mode 100644 index 00000000000..73595cd1caa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressClassSpec.libsonnet @@ -0,0 +1,17 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressClassSpec', url='', help='IngressClassSpec provides information about the class of an Ingress.'), + '#parameters':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + parameters: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { parameters+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { parameters+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { parameters+: { name: name } }, + }, + '#withController':: d.fn(help='Controller refers to the name of the controller that should handle this class. This allows for different "flavors" that are controlled by the same controller. For example, you may have different Parameters for the same implementing controller. This should be specified as a domain-prefixed path no more than 250 characters in length, e.g. "acme.io/ingress-controller". This field is immutable.', args=[d.arg(name='controller', type=d.T.string)]), + withController(controller): { controller: controller }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressRule.libsonnet new file mode 100644 index 00000000000..3a0a5dca509 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressRule.libsonnet @@ -0,0 +1,15 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressRule', url='', help='IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue.'), + '#http':: d.obj(help="HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'."), + http: { + '#withPaths':: d.fn(help='A collection of paths that map requests to backends.', args=[d.arg(name='paths', type=d.T.array)]), + withPaths(paths): { http+: { paths: if std.isArray(v=paths) then paths else [paths] } }, + '#withPathsMixin':: d.fn(help='A collection of paths that map requests to backends.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='paths', type=d.T.array)]), + withPathsMixin(paths): { http+: { paths+: if std.isArray(v=paths) then paths else [paths] } }, + }, + '#withHost':: d.fn(help="Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the 'host' part of the URI as defined in RFC 3986: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to\n the IP in the Spec of the parent Ingress.\n2. The `:` delimiter is not respected because ports are not allowed.\n\t Currently the port of an Ingress is implicitly :80 for http and\n\t :443 for https.\nBoth these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue.\n\nHost can be 'precise' which is a domain name without the terminating dot of a network host (e.g. 'foo.bar.com') or 'wildcard', which is a domain name prefixed with a single wildcard label (e.g. '*.foo.com'). The wildcard character '*' must appear by itself as the first DNS label and matches only a single label. You cannot have a wildcard label by itself (e.g. Host == '*'). Requests will be matched against the Host field in the following way: 1. If Host is precise, the request matches this rule if the http host header is equal to Host. 2. If Host is a wildcard, then the request matches this rule if the http host header is to equal to the suffix (removing the first label) of the wildcard rule.", args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { host: host }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressSpec.libsonnet new file mode 100644 index 00000000000..a61df751c89 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressSpec.libsonnet @@ -0,0 +1,32 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressSpec', url='', help='IngressSpec describes the Ingress the user wishes to exist.'), + '#backend':: d.obj(help='IngressBackend describes all endpoints for a given service and port.'), + backend: { + '#resource':: d.obj(help='TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace.'), + resource: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { backend+: { resource+: { apiGroup: apiGroup } } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { backend+: { resource+: { kind: kind } } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { backend+: { resource+: { name: name } } }, + }, + '#withServiceName':: d.fn(help='Specifies the name of the referenced service.', args=[d.arg(name='serviceName', type=d.T.string)]), + withServiceName(serviceName): { backend+: { serviceName: serviceName } }, + '#withServicePort':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='servicePort', type=d.T.string)]), + withServicePort(servicePort): { backend+: { servicePort: servicePort } }, + }, + '#withIngressClassName':: d.fn(help='IngressClassName is the name of the IngressClass cluster resource. The associated IngressClass defines which controller will implement the resource. This replaces the deprecated `kubernetes.io/ingress.class` annotation. For backwards compatibility, when that annotation is set, it must be given precedence over this field. The controller may emit a warning if the field and annotation have different values. Implementations of this API should ignore Ingresses without a class specified. An IngressClass resource may be marked as default, which can be used to set a default value for this field. For more information, refer to the IngressClass documentation.', args=[d.arg(name='ingressClassName', type=d.T.string)]), + withIngressClassName(ingressClassName): { ingressClassName: ingressClassName }, + '#withRules':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#withTls':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.', args=[d.arg(name='tls', type=d.T.array)]), + withTls(tls): { tls: if std.isArray(v=tls) then tls else [tls] }, + '#withTlsMixin':: d.fn(help='TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tls', type=d.T.array)]), + withTlsMixin(tls): { tls+: if std.isArray(v=tls) then tls else [tls] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressStatus.libsonnet new file mode 100644 index 00000000000..697b5b33c1a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressStatus.libsonnet @@ -0,0 +1,13 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressStatus', url='', help='IngressStatus describe the current state of the Ingress.'), + '#loadBalancer':: d.obj(help='LoadBalancerStatus represents the status of a load-balancer.'), + loadBalancer: { + '#withIngress':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.', args=[d.arg(name='ingress', type=d.T.array)]), + withIngress(ingress): { loadBalancer+: { ingress: if std.isArray(v=ingress) then ingress else [ingress] } }, + '#withIngressMixin':: d.fn(help='Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ingress', type=d.T.array)]), + withIngressMixin(ingress): { loadBalancer+: { ingress+: if std.isArray(v=ingress) then ingress else [ingress] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressTLS.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressTLS.libsonnet new file mode 100644 index 00000000000..fbbe8558c31 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/ingressTLS.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='ingressTLS', url='', help='IngressTLS describes the transport layer security associated with an Ingress.'), + '#withHosts':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.', args=[d.arg(name='hosts', type=d.T.array)]), + withHosts(hosts): { hosts: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withHostsMixin':: d.fn(help='Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hosts', type=d.T.array)]), + withHostsMixin(hosts): { hosts+: if std.isArray(v=hosts) then hosts else [hosts] }, + '#withSecretName':: d.fn(help='SecretName is the name of the secret used to terminate TLS traffic on port 443. Field is left optional to allow TLS routing based on SNI hostname alone. If the SNI host in a listener conflicts with the "Host" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing.', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secretName: secretName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/main.libsonnet new file mode 100644 index 00000000000..5958b0e32bc --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/networking/v1beta1/main.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + httpIngressPath: (import 'httpIngressPath.libsonnet'), + httpIngressRuleValue: (import 'httpIngressRuleValue.libsonnet'), + ingress: (import 'ingress.libsonnet'), + ingressBackend: (import 'ingressBackend.libsonnet'), + ingressClass: (import 'ingressClass.libsonnet'), + ingressClassSpec: (import 'ingressClassSpec.libsonnet'), + ingressRule: (import 'ingressRule.libsonnet'), + ingressSpec: (import 'ingressSpec.libsonnet'), + ingressStatus: (import 'ingressStatus.libsonnet'), + ingressTLS: (import 'ingressTLS.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/main.libsonnet new file mode 100644 index 00000000000..d5e64b0b66e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='node', url='', help=''), + v1alpha1: (import 'v1alpha1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..fc1ded9abfd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/main.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + overhead: (import 'overhead.libsonnet'), + runtimeClass: (import 'runtimeClass.libsonnet'), + runtimeClassSpec: (import 'runtimeClassSpec.libsonnet'), + scheduling: (import 'scheduling.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/overhead.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/overhead.libsonnet new file mode 100644 index 00000000000..768cb90d25c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/overhead.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='overhead', url='', help='Overhead structure represents the resource overhead associated with running a pod.'), + '#withPodFixed':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixed(podFixed): { podFixed: podFixed }, + '#withPodFixedMixin':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixedMixin(podFixed): { podFixed+: podFixed }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClass.libsonnet new file mode 100644 index 00000000000..49c40abfedf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClass.libsonnet @@ -0,0 +1,79 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runtimeClass', url='', help='RuntimeClass defines a class of container runtime supported in the cluster. The RuntimeClass is used to determine which container runtime is used to run all containers in a pod. RuntimeClasses are (currently) manually defined by a user or cluster provisioner, and referenced in the PodSpec. The Kubelet is responsible for resolving the RuntimeClassName reference before running the pod. For more details, see https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Runtimeclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'node.k8s.io/v1alpha1', + kind: 'RuntimeClass', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='RuntimeClassSpec is a specification of a RuntimeClass. It contains parameters that are required to describe the RuntimeClass to the Container Runtime Interface (CRI) implementation, as well as any other components that need to understand how the pod will be run. The RuntimeClassSpec is immutable.'), + spec: { + '#overhead':: d.obj(help='Overhead structure represents the resource overhead associated with running a pod.'), + overhead: { + '#withPodFixed':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixed(podFixed): { spec+: { overhead+: { podFixed: podFixed } } }, + '#withPodFixedMixin':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixedMixin(podFixed): { spec+: { overhead+: { podFixed+: podFixed } } }, + }, + '#scheduling':: d.obj(help='Scheduling specifies the scheduling constraints for nodes supporting a RuntimeClass.'), + scheduling: { + '#withNodeSelector':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { scheduling+: { nodeSelector: nodeSelector } } }, + '#withNodeSelectorMixin':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { scheduling+: { nodeSelector+: nodeSelector } } }, + '#withTolerations':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { scheduling+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + '#withTolerationsMixin':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { scheduling+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } }, + }, + '#withRuntimeHandler':: d.fn(help='RuntimeHandler specifies the underlying runtime and configuration that the CRI implementation will use to handle pods of this class. The possible values are specific to the node & CRI configuration. It is assumed that all handlers are available on every node, and handlers of the same name are equivalent on every node. For example, a handler called "runc" might specify that the runc OCI runtime (using native Linux containers) will be used to run the containers in a pod. The RuntimeHandler must conform to the DNS Label (RFC 1123) requirements and is immutable.', args=[d.arg(name='runtimeHandler', type=d.T.string)]), + withRuntimeHandler(runtimeHandler): { spec+: { runtimeHandler: runtimeHandler } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClassSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClassSpec.libsonnet new file mode 100644 index 00000000000..125a05c9f2a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/runtimeClassSpec.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runtimeClassSpec', url='', help='RuntimeClassSpec is a specification of a RuntimeClass. It contains parameters that are required to describe the RuntimeClass to the Container Runtime Interface (CRI) implementation, as well as any other components that need to understand how the pod will be run. The RuntimeClassSpec is immutable.'), + '#overhead':: d.obj(help='Overhead structure represents the resource overhead associated with running a pod.'), + overhead: { + '#withPodFixed':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixed(podFixed): { overhead+: { podFixed: podFixed } }, + '#withPodFixedMixin':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixedMixin(podFixed): { overhead+: { podFixed+: podFixed } }, + }, + '#scheduling':: d.obj(help='Scheduling specifies the scheduling constraints for nodes supporting a RuntimeClass.'), + scheduling: { + '#withNodeSelector':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { scheduling+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { scheduling+: { nodeSelector+: nodeSelector } }, + '#withTolerations':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { scheduling+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTolerationsMixin':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { scheduling+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + }, + '#withRuntimeHandler':: d.fn(help='RuntimeHandler specifies the underlying runtime and configuration that the CRI implementation will use to handle pods of this class. The possible values are specific to the node & CRI configuration. It is assumed that all handlers are available on every node, and handlers of the same name are equivalent on every node. For example, a handler called "runc" might specify that the runc OCI runtime (using native Linux containers) will be used to run the containers in a pod. The RuntimeHandler must conform to the DNS Label (RFC 1123) requirements and is immutable.', args=[d.arg(name='runtimeHandler', type=d.T.string)]), + withRuntimeHandler(runtimeHandler): { runtimeHandler: runtimeHandler }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/scheduling.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/scheduling.libsonnet new file mode 100644 index 00000000000..0443f726b23 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1alpha1/scheduling.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scheduling', url='', help='Scheduling specifies the scheduling constraints for nodes supporting a RuntimeClass.'), + '#withNodeSelector':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { nodeSelector: nodeSelector }, + '#withNodeSelectorMixin':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { nodeSelector+: nodeSelector }, + '#withTolerations':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#withTolerationsMixin':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/main.libsonnet new file mode 100644 index 00000000000..cbc6521678c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + overhead: (import 'overhead.libsonnet'), + runtimeClass: (import 'runtimeClass.libsonnet'), + scheduling: (import 'scheduling.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/overhead.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/overhead.libsonnet new file mode 100644 index 00000000000..768cb90d25c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/overhead.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='overhead', url='', help='Overhead structure represents the resource overhead associated with running a pod.'), + '#withPodFixed':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixed(podFixed): { podFixed: podFixed }, + '#withPodFixedMixin':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixedMixin(podFixed): { podFixed+: podFixed }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/runtimeClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/runtimeClass.libsonnet new file mode 100644 index 00000000000..b6391946d85 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/runtimeClass.libsonnet @@ -0,0 +1,76 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runtimeClass', url='', help='RuntimeClass defines a class of container runtime supported in the cluster. The RuntimeClass is used to determine which container runtime is used to run all containers in a pod. RuntimeClasses are (currently) manually defined by a user or cluster provisioner, and referenced in the PodSpec. The Kubelet is responsible for resolving the RuntimeClassName reference before running the pod. For more details, see https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Runtimeclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'node.k8s.io/v1beta1', + kind: 'RuntimeClass', + } + self.metadata.withName(name=name), + '#overhead':: d.obj(help='Overhead structure represents the resource overhead associated with running a pod.'), + overhead: { + '#withPodFixed':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixed(podFixed): { overhead+: { podFixed: podFixed } }, + '#withPodFixedMixin':: d.fn(help='PodFixed represents the fixed resource overhead associated with running a pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podFixed', type=d.T.object)]), + withPodFixedMixin(podFixed): { overhead+: { podFixed+: podFixed } }, + }, + '#scheduling':: d.obj(help='Scheduling specifies the scheduling constraints for nodes supporting a RuntimeClass.'), + scheduling: { + '#withNodeSelector':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { scheduling+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { scheduling+: { nodeSelector+: nodeSelector } }, + '#withTolerations':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { scheduling+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTolerationsMixin':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { scheduling+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + }, + '#withHandler':: d.fn(help='Handler specifies the underlying runtime and configuration that the CRI implementation will use to handle pods of this class. The possible values are specific to the node & CRI configuration. It is assumed that all handlers are available on every node, and handlers of the same name are equivalent on every node. For example, a handler called "runc" might specify that the runc OCI runtime (using native Linux containers) will be used to run the containers in a pod. The Handler must conform to the DNS Label (RFC 1123) requirements, and is immutable.', args=[d.arg(name='handler', type=d.T.string)]), + withHandler(handler): { handler: handler }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/scheduling.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/scheduling.libsonnet new file mode 100644 index 00000000000..0443f726b23 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/node/v1beta1/scheduling.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scheduling', url='', help='Scheduling specifies the scheduling constraints for nodes supporting a RuntimeClass.'), + '#withNodeSelector':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { nodeSelector: nodeSelector }, + '#withNodeSelectorMixin':: d.fn(help="nodeSelector lists labels that must be present on nodes that support this RuntimeClass. Pods using this RuntimeClass can only be scheduled to a node matched by this selector. The RuntimeClass nodeSelector is merged with a pod's existing nodeSelector. Any conflicts will cause the pod to be rejected in admission.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { nodeSelector+: nodeSelector }, + '#withTolerations':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#withTolerationsMixin':: d.fn(help='tolerations are appended (excluding duplicates) to pods running with this RuntimeClass during admission, effectively unioning the set of nodes tolerated by the pod and the RuntimeClass.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/main.libsonnet new file mode 100644 index 00000000000..4636818ff2f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='policy', url='', help=''), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedCSIDriver.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedCSIDriver.libsonnet new file mode 100644 index 00000000000..f8e7dadc0a2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedCSIDriver.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='allowedCSIDriver', url='', help='AllowedCSIDriver represents a single inline CSI Driver that is allowed to be used.'), + '#withName':: d.fn(help='Name is the registered name of the CSI driver', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedFlexVolume.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedFlexVolume.libsonnet new file mode 100644 index 00000000000..9d5ce68dbcb --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedFlexVolume.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='allowedFlexVolume', url='', help='AllowedFlexVolume represents a single Flexvolume that is allowed to be used.'), + '#withDriver':: d.fn(help='driver is the name of the Flexvolume driver.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { driver: driver }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedHostPath.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedHostPath.libsonnet new file mode 100644 index 00000000000..8daf70220e0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/allowedHostPath.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='allowedHostPath', url='', help='AllowedHostPath defines the host volume conditions that will be enabled by a policy for pods to use. It requires the path prefix to be defined.'), + '#withPathPrefix':: d.fn(help='pathPrefix is the path prefix that the host volume must match. It does not support `*`. Trailing slashes are trimmed when validating the path prefix with a host path.\n\nExamples: `/foo` would allow `/foo`, `/foo/` and `/foo/bar` `/foo` would not allow `/food` or `/etc/foo`', args=[d.arg(name='pathPrefix', type=d.T.string)]), + withPathPrefix(pathPrefix): { pathPrefix: pathPrefix }, + '#withReadOnly':: d.fn(help='when set to true, will allow host volumes matching the pathPrefix only if all volume mounts are readOnly.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/eviction.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/eviction.libsonnet new file mode 100644 index 00000000000..c0bf81d9985 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/eviction.libsonnet @@ -0,0 +1,80 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='eviction', url='', help='Eviction evicts a pod from its node subject to certain policies and safety constraints. This is a subresource of Pod. A request to cause such an eviction is created by POSTing to .../pods//evictions.'), + '#deleteOptions':: d.obj(help='DeleteOptions may be provided when deleting an API object.'), + deleteOptions: { + '#preconditions':: d.obj(help='Preconditions must be fulfilled before an operation (update, delete, etc.) is carried out.'), + preconditions: { + '#withResourceVersion':: d.fn(help='Specifies the target ResourceVersion', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { deleteOptions+: { preconditions+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='Specifies the target UID.', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { deleteOptions+: { preconditions+: { uid: uid } } }, + }, + '#withApiVersion':: d.fn(help='APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { deleteOptions+: { apiVersion: apiVersion } }, + '#withDryRun':: d.fn(help='When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed', args=[d.arg(name='dryRun', type=d.T.array)]), + withDryRun(dryRun): { deleteOptions+: { dryRun: if std.isArray(v=dryRun) then dryRun else [dryRun] } }, + '#withDryRunMixin':: d.fn(help='When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='dryRun', type=d.T.array)]), + withDryRunMixin(dryRun): { deleteOptions+: { dryRun+: if std.isArray(v=dryRun) then dryRun else [dryRun] } }, + '#withGracePeriodSeconds':: d.fn(help='The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately.', args=[d.arg(name='gracePeriodSeconds', type=d.T.integer)]), + withGracePeriodSeconds(gracePeriodSeconds): { deleteOptions+: { gracePeriodSeconds: gracePeriodSeconds } }, + '#withKind':: d.fn(help='Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { deleteOptions+: { kind: kind } }, + '#withOrphanDependents':: d.fn(help="Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the 'orphan' finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both.", args=[d.arg(name='orphanDependents', type=d.T.boolean)]), + withOrphanDependents(orphanDependents): { deleteOptions+: { orphanDependents: orphanDependents } }, + '#withPropagationPolicy':: d.fn(help="Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground.", args=[d.arg(name='propagationPolicy', type=d.T.string)]), + withPropagationPolicy(propagationPolicy): { deleteOptions+: { propagationPolicy: propagationPolicy } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Eviction', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'policy/v1beta1', + kind: 'Eviction', + } + self.metadata.withName(name=name), + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/fsGroupStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/fsGroupStrategyOptions.libsonnet new file mode 100644 index 00000000000..9ff2b429fce --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/fsGroupStrategyOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='fsGroupStrategyOptions', url='', help='FSGroupStrategyOptions defines the strategy type and options used to create the strategy.'), + '#withRanges':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { ranges: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { ranges+: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what FSGroup is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { rule: rule }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/hostPortRange.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/hostPortRange.libsonnet new file mode 100644 index 00000000000..777c3f7c8ed --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/hostPortRange.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='hostPortRange', url='', help='HostPortRange defines a range of host ports that will be enabled by a policy for pods to use. It requires both the start and end to be defined.'), + '#withMax':: d.fn(help='max is the end of the range, inclusive.', args=[d.arg(name='max', type=d.T.integer)]), + withMax(max): { max: max }, + '#withMin':: d.fn(help='min is the start of the range, inclusive.', args=[d.arg(name='min', type=d.T.integer)]), + withMin(min): { min: min }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/idRange.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/idRange.libsonnet new file mode 100644 index 00000000000..30e7e28bd65 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/idRange.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='idRange', url='', help='IDRange provides a min/max of an allowed range of IDs.'), + '#withMax':: d.fn(help='max is the end of the range, inclusive.', args=[d.arg(name='max', type=d.T.integer)]), + withMax(max): { max: max }, + '#withMin':: d.fn(help='min is the start of the range, inclusive.', args=[d.arg(name='min', type=d.T.integer)]), + withMin(min): { min: min }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/main.libsonnet new file mode 100644 index 00000000000..3b0bb525f0f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/main.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + allowedCSIDriver: (import 'allowedCSIDriver.libsonnet'), + allowedFlexVolume: (import 'allowedFlexVolume.libsonnet'), + allowedHostPath: (import 'allowedHostPath.libsonnet'), + eviction: (import 'eviction.libsonnet'), + fsGroupStrategyOptions: (import 'fsGroupStrategyOptions.libsonnet'), + hostPortRange: (import 'hostPortRange.libsonnet'), + idRange: (import 'idRange.libsonnet'), + podDisruptionBudget: (import 'podDisruptionBudget.libsonnet'), + podDisruptionBudgetSpec: (import 'podDisruptionBudgetSpec.libsonnet'), + podDisruptionBudgetStatus: (import 'podDisruptionBudgetStatus.libsonnet'), + podSecurityPolicy: (import 'podSecurityPolicy.libsonnet'), + podSecurityPolicySpec: (import 'podSecurityPolicySpec.libsonnet'), + runAsGroupStrategyOptions: (import 'runAsGroupStrategyOptions.libsonnet'), + runAsUserStrategyOptions: (import 'runAsUserStrategyOptions.libsonnet'), + runtimeClassStrategyOptions: (import 'runtimeClassStrategyOptions.libsonnet'), + seLinuxStrategyOptions: (import 'seLinuxStrategyOptions.libsonnet'), + supplementalGroupsStrategyOptions: (import 'supplementalGroupsStrategyOptions.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudget.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudget.libsonnet new file mode 100644 index 00000000000..009b193d67a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudget.libsonnet @@ -0,0 +1,74 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podDisruptionBudget', url='', help='PodDisruptionBudget is an object to define the max disruption that can be caused to a collection of pods'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Poddisruptionbudget', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'policy/v1beta1', + kind: 'PodDisruptionBudget', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PodDisruptionBudgetSpec is a description of a PodDisruptionBudget.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { spec+: { maxUnavailable: maxUnavailable } }, + '#withMinAvailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='minAvailable', type=d.T.string)]), + withMinAvailable(minAvailable): { spec+: { minAvailable: minAvailable } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetSpec.libsonnet new file mode 100644 index 00000000000..d7c9a5ea726 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetSpec.libsonnet @@ -0,0 +1,21 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podDisruptionBudgetSpec', url='', help='PodDisruptionBudgetSpec is a description of a PodDisruptionBudget.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withMaxUnavailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='maxUnavailable', type=d.T.string)]), + withMaxUnavailable(maxUnavailable): { maxUnavailable: maxUnavailable }, + '#withMinAvailable':: d.fn(help='IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number.', args=[d.arg(name='minAvailable', type=d.T.string)]), + withMinAvailable(minAvailable): { minAvailable: minAvailable }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetStatus.libsonnet new file mode 100644 index 00000000000..7b3e66c2ded --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podDisruptionBudgetStatus.libsonnet @@ -0,0 +1,20 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podDisruptionBudgetStatus', url='', help='PodDisruptionBudgetStatus represents information about the status of a PodDisruptionBudget. Status may trail the actual state of a system.'), + '#withCurrentHealthy':: d.fn(help='current number of healthy pods', args=[d.arg(name='currentHealthy', type=d.T.integer)]), + withCurrentHealthy(currentHealthy): { currentHealthy: currentHealthy }, + '#withDesiredHealthy':: d.fn(help='minimum desired number of healthy pods', args=[d.arg(name='desiredHealthy', type=d.T.integer)]), + withDesiredHealthy(desiredHealthy): { desiredHealthy: desiredHealthy }, + '#withDisruptedPods':: d.fn(help="DisruptedPods contains information about pods whose eviction was processed by the API server eviction subresource handler but has not yet been observed by the PodDisruptionBudget controller. A pod will be in this map from the time when the API server processed the eviction request to the time when the pod is seen by PDB controller as having been marked for deletion (or after a timeout). The key in the map is the name of the pod and the value is the time when the API server processed the eviction request. If the deletion didn't occur and a pod is still there it will be removed from the list automatically by PodDisruptionBudget controller after some time. If everything goes smooth this map should be empty for the most of the time. Large number of entries in the map may indicate problems with pod deletions.", args=[d.arg(name='disruptedPods', type=d.T.object)]), + withDisruptedPods(disruptedPods): { disruptedPods: disruptedPods }, + '#withDisruptedPodsMixin':: d.fn(help="DisruptedPods contains information about pods whose eviction was processed by the API server eviction subresource handler but has not yet been observed by the PodDisruptionBudget controller. A pod will be in this map from the time when the API server processed the eviction request to the time when the pod is seen by PDB controller as having been marked for deletion (or after a timeout). The key in the map is the name of the pod and the value is the time when the API server processed the eviction request. If the deletion didn't occur and a pod is still there it will be removed from the list automatically by PodDisruptionBudget controller after some time. If everything goes smooth this map should be empty for the most of the time. Large number of entries in the map may indicate problems with pod deletions.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='disruptedPods', type=d.T.object)]), + withDisruptedPodsMixin(disruptedPods): { disruptedPods+: disruptedPods }, + '#withDisruptionsAllowed':: d.fn(help='Number of pod disruptions that are currently allowed.', args=[d.arg(name='disruptionsAllowed', type=d.T.integer)]), + withDisruptionsAllowed(disruptionsAllowed): { disruptionsAllowed: disruptionsAllowed }, + '#withExpectedPods':: d.fn(help='total number of pods counted by this disruption budget', args=[d.arg(name='expectedPods', type=d.T.integer)]), + withExpectedPods(expectedPods): { expectedPods: expectedPods }, + '#withObservedGeneration':: d.fn(help="Most recent generation observed when updating this PDB status. DisruptionsAllowed and other status information is valid only if observedGeneration equals to PDB's object generation.", args=[d.arg(name='observedGeneration', type=d.T.integer)]), + withObservedGeneration(observedGeneration): { observedGeneration: observedGeneration }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicy.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicy.libsonnet new file mode 100644 index 00000000000..15c0f9889e8 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicy.libsonnet @@ -0,0 +1,178 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podSecurityPolicy', url='', help='PodSecurityPolicy governs the ability to make requests that affect the Security Context that will be applied to a pod and container.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Podsecuritypolicy', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'policy/v1beta1', + kind: 'PodSecurityPolicy', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PodSecurityPolicySpec defines the policy enforced.'), + spec: { + '#fsGroup':: d.obj(help='FSGroupStrategyOptions defines the strategy type and options used to create the strategy.'), + fsGroup: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { spec+: { fsGroup+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { spec+: { fsGroup+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what FSGroup is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { spec+: { fsGroup+: { rule: rule } } }, + }, + '#runAsGroup':: d.obj(help='RunAsGroupStrategyOptions defines the strategy type and any options used to create the strategy.'), + runAsGroup: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { spec+: { runAsGroup+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { spec+: { runAsGroup+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsGroup values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { spec+: { runAsGroup+: { rule: rule } } }, + }, + '#runAsUser':: d.obj(help='RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy.'), + runAsUser: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { spec+: { runAsUser+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { spec+: { runAsUser+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsUser values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { spec+: { runAsUser+: { rule: rule } } }, + }, + '#runtimeClass':: d.obj(help='RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses for a pod.'), + runtimeClass: { + '#withAllowedRuntimeClassNames':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNames(allowedRuntimeClassNames): { spec+: { runtimeClass+: { allowedRuntimeClassNames: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] } } }, + '#withAllowedRuntimeClassNamesMixin':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNamesMixin(allowedRuntimeClassNames): { spec+: { runtimeClass+: { allowedRuntimeClassNames+: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] } } }, + '#withDefaultRuntimeClassName':: d.fn(help='defaultRuntimeClassName is the default RuntimeClassName to set on the pod. The default MUST be allowed by the allowedRuntimeClassNames list. A value of nil does not mutate the Pod.', args=[d.arg(name='defaultRuntimeClassName', type=d.T.string)]), + withDefaultRuntimeClassName(defaultRuntimeClassName): { spec+: { runtimeClass+: { defaultRuntimeClassName: defaultRuntimeClassName } } }, + }, + '#seLinux':: d.obj(help='SELinuxStrategyOptions defines the strategy type and any options used to create the strategy.'), + seLinux: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { seLinux+: { seLinuxOptions+: { level: level } } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { seLinux+: { seLinuxOptions+: { role: role } } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { seLinux+: { seLinuxOptions+: { type: type } } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { seLinux+: { seLinuxOptions+: { user: user } } } }, + }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable labels that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { spec+: { seLinux+: { rule: rule } } }, + }, + '#supplementalGroups':: d.obj(help='SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy.'), + supplementalGroups: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { spec+: { supplementalGroups+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { spec+: { supplementalGroups+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what supplemental groups is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { spec+: { supplementalGroups+: { rule: rule } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true.', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { spec+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withAllowedCSIDrivers':: d.fn(help='AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate.', args=[d.arg(name='allowedCSIDrivers', type=d.T.array)]), + withAllowedCSIDrivers(allowedCSIDrivers): { spec+: { allowedCSIDrivers: if std.isArray(v=allowedCSIDrivers) then allowedCSIDrivers else [allowedCSIDrivers] } }, + '#withAllowedCSIDriversMixin':: d.fn(help='AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedCSIDrivers', type=d.T.array)]), + withAllowedCSIDriversMixin(allowedCSIDrivers): { spec+: { allowedCSIDrivers+: if std.isArray(v=allowedCSIDrivers) then allowedCSIDrivers else [allowedCSIDrivers] } }, + '#withAllowedCapabilities':: d.fn(help="allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.", args=[d.arg(name='allowedCapabilities', type=d.T.array)]), + withAllowedCapabilities(allowedCapabilities): { spec+: { allowedCapabilities: if std.isArray(v=allowedCapabilities) then allowedCapabilities else [allowedCapabilities] } }, + '#withAllowedCapabilitiesMixin':: d.fn(help="allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='allowedCapabilities', type=d.T.array)]), + withAllowedCapabilitiesMixin(allowedCapabilities): { spec+: { allowedCapabilities+: if std.isArray(v=allowedCapabilities) then allowedCapabilities else [allowedCapabilities] } }, + '#withAllowedFlexVolumes':: d.fn(help='allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field.', args=[d.arg(name='allowedFlexVolumes', type=d.T.array)]), + withAllowedFlexVolumes(allowedFlexVolumes): { spec+: { allowedFlexVolumes: if std.isArray(v=allowedFlexVolumes) then allowedFlexVolumes else [allowedFlexVolumes] } }, + '#withAllowedFlexVolumesMixin':: d.fn(help='allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedFlexVolumes', type=d.T.array)]), + withAllowedFlexVolumesMixin(allowedFlexVolumes): { spec+: { allowedFlexVolumes+: if std.isArray(v=allowedFlexVolumes) then allowedFlexVolumes else [allowedFlexVolumes] } }, + '#withAllowedHostPaths':: d.fn(help='allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.', args=[d.arg(name='allowedHostPaths', type=d.T.array)]), + withAllowedHostPaths(allowedHostPaths): { spec+: { allowedHostPaths: if std.isArray(v=allowedHostPaths) then allowedHostPaths else [allowedHostPaths] } }, + '#withAllowedHostPathsMixin':: d.fn(help='allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedHostPaths', type=d.T.array)]), + withAllowedHostPathsMixin(allowedHostPaths): { spec+: { allowedHostPaths+: if std.isArray(v=allowedHostPaths) then allowedHostPaths else [allowedHostPaths] } }, + '#withAllowedProcMountTypes':: d.fn(help='AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.', args=[d.arg(name='allowedProcMountTypes', type=d.T.array)]), + withAllowedProcMountTypes(allowedProcMountTypes): { spec+: { allowedProcMountTypes: if std.isArray(v=allowedProcMountTypes) then allowedProcMountTypes else [allowedProcMountTypes] } }, + '#withAllowedProcMountTypesMixin':: d.fn(help='AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedProcMountTypes', type=d.T.array)]), + withAllowedProcMountTypesMixin(allowedProcMountTypes): { spec+: { allowedProcMountTypes+: if std.isArray(v=allowedProcMountTypes) then allowedProcMountTypes else [allowedProcMountTypes] } }, + '#withAllowedUnsafeSysctls':: d.fn(help='allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc.', args=[d.arg(name='allowedUnsafeSysctls', type=d.T.array)]), + withAllowedUnsafeSysctls(allowedUnsafeSysctls): { spec+: { allowedUnsafeSysctls: if std.isArray(v=allowedUnsafeSysctls) then allowedUnsafeSysctls else [allowedUnsafeSysctls] } }, + '#withAllowedUnsafeSysctlsMixin':: d.fn(help='allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedUnsafeSysctls', type=d.T.array)]), + withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls): { spec+: { allowedUnsafeSysctls+: if std.isArray(v=allowedUnsafeSysctls) then allowedUnsafeSysctls else [allowedUnsafeSysctls] } }, + '#withDefaultAddCapabilities':: d.fn(help='defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.', args=[d.arg(name='defaultAddCapabilities', type=d.T.array)]), + withDefaultAddCapabilities(defaultAddCapabilities): { spec+: { defaultAddCapabilities: if std.isArray(v=defaultAddCapabilities) then defaultAddCapabilities else [defaultAddCapabilities] } }, + '#withDefaultAddCapabilitiesMixin':: d.fn(help='defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='defaultAddCapabilities', type=d.T.array)]), + withDefaultAddCapabilitiesMixin(defaultAddCapabilities): { spec+: { defaultAddCapabilities+: if std.isArray(v=defaultAddCapabilities) then defaultAddCapabilities else [defaultAddCapabilities] } }, + '#withDefaultAllowPrivilegeEscalation':: d.fn(help='defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process.', args=[d.arg(name='defaultAllowPrivilegeEscalation', type=d.T.boolean)]), + withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation): { spec+: { defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation } }, + '#withForbiddenSysctls':: d.fn(help='forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc.', args=[d.arg(name='forbiddenSysctls', type=d.T.array)]), + withForbiddenSysctls(forbiddenSysctls): { spec+: { forbiddenSysctls: if std.isArray(v=forbiddenSysctls) then forbiddenSysctls else [forbiddenSysctls] } }, + '#withForbiddenSysctlsMixin':: d.fn(help='forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='forbiddenSysctls', type=d.T.array)]), + withForbiddenSysctlsMixin(forbiddenSysctls): { spec+: { forbiddenSysctls+: if std.isArray(v=forbiddenSysctls) then forbiddenSysctls else [forbiddenSysctls] } }, + '#withHostIPC':: d.fn(help='hostIPC determines if the policy allows the use of HostIPC in the pod spec.', args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { spec+: { hostIPC: hostIPC } }, + '#withHostNetwork':: d.fn(help='hostNetwork determines if the policy allows the use of HostNetwork in the pod spec.', args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { hostNetwork: hostNetwork } }, + '#withHostPID':: d.fn(help='hostPID determines if the policy allows the use of HostPID in the pod spec.', args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { spec+: { hostPID: hostPID } }, + '#withHostPorts':: d.fn(help='hostPorts determines which host port ranges are allowed to be exposed.', args=[d.arg(name='hostPorts', type=d.T.array)]), + withHostPorts(hostPorts): { spec+: { hostPorts: if std.isArray(v=hostPorts) then hostPorts else [hostPorts] } }, + '#withHostPortsMixin':: d.fn(help='hostPorts determines which host port ranges are allowed to be exposed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostPorts', type=d.T.array)]), + withHostPortsMixin(hostPorts): { spec+: { hostPorts+: if std.isArray(v=hostPorts) then hostPorts else [hostPorts] } }, + '#withPrivileged':: d.fn(help='privileged determines if a pod can request to be run as privileged.', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { spec+: { privileged: privileged } }, + '#withReadOnlyRootFilesystem':: d.fn(help='readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to.', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { spec+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRequiredDropCapabilities':: d.fn(help='requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.', args=[d.arg(name='requiredDropCapabilities', type=d.T.array)]), + withRequiredDropCapabilities(requiredDropCapabilities): { spec+: { requiredDropCapabilities: if std.isArray(v=requiredDropCapabilities) then requiredDropCapabilities else [requiredDropCapabilities] } }, + '#withRequiredDropCapabilitiesMixin':: d.fn(help='requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDropCapabilities', type=d.T.array)]), + withRequiredDropCapabilitiesMixin(requiredDropCapabilities): { spec+: { requiredDropCapabilities+: if std.isArray(v=requiredDropCapabilities) then requiredDropCapabilities else [requiredDropCapabilities] } }, + '#withVolumes':: d.fn(help="volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } }, + '#withVolumesMixin':: d.fn(help="volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicySpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicySpec.libsonnet new file mode 100644 index 00000000000..3d5969b897b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/podSecurityPolicySpec.libsonnet @@ -0,0 +1,125 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podSecurityPolicySpec', url='', help='PodSecurityPolicySpec defines the policy enforced.'), + '#fsGroup':: d.obj(help='FSGroupStrategyOptions defines the strategy type and options used to create the strategy.'), + fsGroup: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { fsGroup+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { fsGroup+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what FSGroup is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { fsGroup+: { rule: rule } }, + }, + '#runAsGroup':: d.obj(help='RunAsGroupStrategyOptions defines the strategy type and any options used to create the strategy.'), + runAsGroup: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { runAsGroup+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { runAsGroup+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsGroup values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { runAsGroup+: { rule: rule } }, + }, + '#runAsUser':: d.obj(help='RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy.'), + runAsUser: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { runAsUser+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { runAsUser+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsUser values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { runAsUser+: { rule: rule } }, + }, + '#runtimeClass':: d.obj(help='RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses for a pod.'), + runtimeClass: { + '#withAllowedRuntimeClassNames':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNames(allowedRuntimeClassNames): { runtimeClass+: { allowedRuntimeClassNames: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] } }, + '#withAllowedRuntimeClassNamesMixin':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNamesMixin(allowedRuntimeClassNames): { runtimeClass+: { allowedRuntimeClassNames+: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] } }, + '#withDefaultRuntimeClassName':: d.fn(help='defaultRuntimeClassName is the default RuntimeClassName to set on the pod. The default MUST be allowed by the allowedRuntimeClassNames list. A value of nil does not mutate the Pod.', args=[d.arg(name='defaultRuntimeClassName', type=d.T.string)]), + withDefaultRuntimeClassName(defaultRuntimeClassName): { runtimeClass+: { defaultRuntimeClassName: defaultRuntimeClassName } }, + }, + '#seLinux':: d.obj(help='SELinuxStrategyOptions defines the strategy type and any options used to create the strategy.'), + seLinux: { + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { seLinux+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { seLinux+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seLinux+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { seLinux+: { seLinuxOptions+: { user: user } } }, + }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable labels that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { seLinux+: { rule: rule } }, + }, + '#supplementalGroups':: d.obj(help='SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy.'), + supplementalGroups: { + '#withRanges':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { supplementalGroups+: { ranges: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { supplementalGroups+: { ranges+: if std.isArray(v=ranges) then ranges else [ranges] } }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what supplemental groups is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { supplementalGroups+: { rule: rule } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true.', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { allowPrivilegeEscalation: allowPrivilegeEscalation }, + '#withAllowedCSIDrivers':: d.fn(help='AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate.', args=[d.arg(name='allowedCSIDrivers', type=d.T.array)]), + withAllowedCSIDrivers(allowedCSIDrivers): { allowedCSIDrivers: if std.isArray(v=allowedCSIDrivers) then allowedCSIDrivers else [allowedCSIDrivers] }, + '#withAllowedCSIDriversMixin':: d.fn(help='AllowedCSIDrivers is an allowlist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value indicates that any CSI driver can be used for inline ephemeral volumes. This is a beta field, and is only honored if the API server enables the CSIInlineVolume feature gate.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedCSIDrivers', type=d.T.array)]), + withAllowedCSIDriversMixin(allowedCSIDrivers): { allowedCSIDrivers+: if std.isArray(v=allowedCSIDrivers) then allowedCSIDrivers else [allowedCSIDrivers] }, + '#withAllowedCapabilities':: d.fn(help="allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.", args=[d.arg(name='allowedCapabilities', type=d.T.array)]), + withAllowedCapabilities(allowedCapabilities): { allowedCapabilities: if std.isArray(v=allowedCapabilities) then allowedCapabilities else [allowedCapabilities] }, + '#withAllowedCapabilitiesMixin':: d.fn(help="allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='allowedCapabilities', type=d.T.array)]), + withAllowedCapabilitiesMixin(allowedCapabilities): { allowedCapabilities+: if std.isArray(v=allowedCapabilities) then allowedCapabilities else [allowedCapabilities] }, + '#withAllowedFlexVolumes':: d.fn(help='allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field.', args=[d.arg(name='allowedFlexVolumes', type=d.T.array)]), + withAllowedFlexVolumes(allowedFlexVolumes): { allowedFlexVolumes: if std.isArray(v=allowedFlexVolumes) then allowedFlexVolumes else [allowedFlexVolumes] }, + '#withAllowedFlexVolumesMixin':: d.fn(help='allowedFlexVolumes is an allowlist of Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedFlexVolumes', type=d.T.array)]), + withAllowedFlexVolumesMixin(allowedFlexVolumes): { allowedFlexVolumes+: if std.isArray(v=allowedFlexVolumes) then allowedFlexVolumes else [allowedFlexVolumes] }, + '#withAllowedHostPaths':: d.fn(help='allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.', args=[d.arg(name='allowedHostPaths', type=d.T.array)]), + withAllowedHostPaths(allowedHostPaths): { allowedHostPaths: if std.isArray(v=allowedHostPaths) then allowedHostPaths else [allowedHostPaths] }, + '#withAllowedHostPathsMixin':: d.fn(help='allowedHostPaths is an allowlist of host paths. Empty indicates that all host paths may be used.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedHostPaths', type=d.T.array)]), + withAllowedHostPathsMixin(allowedHostPaths): { allowedHostPaths+: if std.isArray(v=allowedHostPaths) then allowedHostPaths else [allowedHostPaths] }, + '#withAllowedProcMountTypes':: d.fn(help='AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.', args=[d.arg(name='allowedProcMountTypes', type=d.T.array)]), + withAllowedProcMountTypes(allowedProcMountTypes): { allowedProcMountTypes: if std.isArray(v=allowedProcMountTypes) then allowedProcMountTypes else [allowedProcMountTypes] }, + '#withAllowedProcMountTypesMixin':: d.fn(help='AllowedProcMountTypes is an allowlist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedProcMountTypes', type=d.T.array)]), + withAllowedProcMountTypesMixin(allowedProcMountTypes): { allowedProcMountTypes+: if std.isArray(v=allowedProcMountTypes) then allowedProcMountTypes else [allowedProcMountTypes] }, + '#withAllowedUnsafeSysctls':: d.fn(help='allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc.', args=[d.arg(name='allowedUnsafeSysctls', type=d.T.array)]), + withAllowedUnsafeSysctls(allowedUnsafeSysctls): { allowedUnsafeSysctls: if std.isArray(v=allowedUnsafeSysctls) then allowedUnsafeSysctls else [allowedUnsafeSysctls] }, + '#withAllowedUnsafeSysctlsMixin':: d.fn(help='allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to allowlist all allowed unsafe sysctls explicitly to avoid rejection.\n\nExamples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedUnsafeSysctls', type=d.T.array)]), + withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls): { allowedUnsafeSysctls+: if std.isArray(v=allowedUnsafeSysctls) then allowedUnsafeSysctls else [allowedUnsafeSysctls] }, + '#withDefaultAddCapabilities':: d.fn(help='defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.', args=[d.arg(name='defaultAddCapabilities', type=d.T.array)]), + withDefaultAddCapabilities(defaultAddCapabilities): { defaultAddCapabilities: if std.isArray(v=defaultAddCapabilities) then defaultAddCapabilities else [defaultAddCapabilities] }, + '#withDefaultAddCapabilitiesMixin':: d.fn(help='defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='defaultAddCapabilities', type=d.T.array)]), + withDefaultAddCapabilitiesMixin(defaultAddCapabilities): { defaultAddCapabilities+: if std.isArray(v=defaultAddCapabilities) then defaultAddCapabilities else [defaultAddCapabilities] }, + '#withDefaultAllowPrivilegeEscalation':: d.fn(help='defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process.', args=[d.arg(name='defaultAllowPrivilegeEscalation', type=d.T.boolean)]), + withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation): { defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation }, + '#withForbiddenSysctls':: d.fn(help='forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc.', args=[d.arg(name='forbiddenSysctls', type=d.T.array)]), + withForbiddenSysctls(forbiddenSysctls): { forbiddenSysctls: if std.isArray(v=forbiddenSysctls) then forbiddenSysctls else [forbiddenSysctls] }, + '#withForbiddenSysctlsMixin':: d.fn(help='forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden.\n\nExamples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='forbiddenSysctls', type=d.T.array)]), + withForbiddenSysctlsMixin(forbiddenSysctls): { forbiddenSysctls+: if std.isArray(v=forbiddenSysctls) then forbiddenSysctls else [forbiddenSysctls] }, + '#withHostIPC':: d.fn(help='hostIPC determines if the policy allows the use of HostIPC in the pod spec.', args=[d.arg(name='hostIPC', type=d.T.boolean)]), + withHostIPC(hostIPC): { hostIPC: hostIPC }, + '#withHostNetwork':: d.fn(help='hostNetwork determines if the policy allows the use of HostNetwork in the pod spec.', args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { hostNetwork: hostNetwork }, + '#withHostPID':: d.fn(help='hostPID determines if the policy allows the use of HostPID in the pod spec.', args=[d.arg(name='hostPID', type=d.T.boolean)]), + withHostPID(hostPID): { hostPID: hostPID }, + '#withHostPorts':: d.fn(help='hostPorts determines which host port ranges are allowed to be exposed.', args=[d.arg(name='hostPorts', type=d.T.array)]), + withHostPorts(hostPorts): { hostPorts: if std.isArray(v=hostPorts) then hostPorts else [hostPorts] }, + '#withHostPortsMixin':: d.fn(help='hostPorts determines which host port ranges are allowed to be exposed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostPorts', type=d.T.array)]), + withHostPortsMixin(hostPorts): { hostPorts+: if std.isArray(v=hostPorts) then hostPorts else [hostPorts] }, + '#withPrivileged':: d.fn(help='privileged determines if a pod can request to be run as privileged.', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { privileged: privileged }, + '#withReadOnlyRootFilesystem':: d.fn(help='readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to.', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { readOnlyRootFilesystem: readOnlyRootFilesystem }, + '#withRequiredDropCapabilities':: d.fn(help='requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.', args=[d.arg(name='requiredDropCapabilities', type=d.T.array)]), + withRequiredDropCapabilities(requiredDropCapabilities): { requiredDropCapabilities: if std.isArray(v=requiredDropCapabilities) then requiredDropCapabilities else [requiredDropCapabilities] }, + '#withRequiredDropCapabilitiesMixin':: d.fn(help='requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDropCapabilities', type=d.T.array)]), + withRequiredDropCapabilitiesMixin(requiredDropCapabilities): { requiredDropCapabilities+: if std.isArray(v=requiredDropCapabilities) then requiredDropCapabilities else [requiredDropCapabilities] }, + '#withVolumes':: d.fn(help="volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.", args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { volumes: if std.isArray(v=volumes) then volumes else [volumes] }, + '#withVolumesMixin':: d.fn(help="volumes is an allowlist of volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { volumes+: if std.isArray(v=volumes) then volumes else [volumes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsGroupStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsGroupStrategyOptions.libsonnet new file mode 100644 index 00000000000..36af206820d --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsGroupStrategyOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runAsGroupStrategyOptions', url='', help='RunAsGroupStrategyOptions defines the strategy type and any options used to create the strategy.'), + '#withRanges':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { ranges: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { ranges+: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsGroup values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { rule: rule }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsUserStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsUserStrategyOptions.libsonnet new file mode 100644 index 00000000000..3a4fbe770fa --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runAsUserStrategyOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runAsUserStrategyOptions', url='', help='RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy.'), + '#withRanges':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { ranges: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { ranges+: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable RunAsUser values that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { rule: rule }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runtimeClassStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runtimeClassStrategyOptions.libsonnet new file mode 100644 index 00000000000..491c0d47c6f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/runtimeClassStrategyOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='runtimeClassStrategyOptions', url='', help='RuntimeClassStrategyOptions define the strategy that will dictate the allowable RuntimeClasses for a pod.'), + '#withAllowedRuntimeClassNames':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNames(allowedRuntimeClassNames): { allowedRuntimeClassNames: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] }, + '#withAllowedRuntimeClassNamesMixin':: d.fn(help='allowedRuntimeClassNames is an allowlist of RuntimeClass names that may be specified on a pod. A value of "*" means that any RuntimeClass name is allowed, and must be the only item in the list. An empty list requires the RuntimeClassName field to be unset.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedRuntimeClassNames', type=d.T.array)]), + withAllowedRuntimeClassNamesMixin(allowedRuntimeClassNames): { allowedRuntimeClassNames+: if std.isArray(v=allowedRuntimeClassNames) then allowedRuntimeClassNames else [allowedRuntimeClassNames] }, + '#withDefaultRuntimeClassName':: d.fn(help='defaultRuntimeClassName is the default RuntimeClassName to set on the pod. The default MUST be allowed by the allowedRuntimeClassNames list. A value of nil does not mutate the Pod.', args=[d.arg(name='defaultRuntimeClassName', type=d.T.string)]), + withDefaultRuntimeClassName(defaultRuntimeClassName): { defaultRuntimeClassName: defaultRuntimeClassName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/seLinuxStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/seLinuxStrategyOptions.libsonnet new file mode 100644 index 00000000000..870a8e85f90 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/seLinuxStrategyOptions.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='seLinuxStrategyOptions', url='', help='SELinuxStrategyOptions defines the strategy type and any options used to create the strategy.'), + '#seLinuxOptions':: d.obj(help='SELinuxOptions are the labels to be applied to the container'), + seLinuxOptions: { + '#withLevel':: d.fn(help='Level is SELinux level label that applies to the container.', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { seLinuxOptions+: { level: level } }, + '#withRole':: d.fn(help='Role is a SELinux role label that applies to the container.', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { seLinuxOptions+: { role: role } }, + '#withType':: d.fn(help='Type is a SELinux type label that applies to the container.', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { seLinuxOptions+: { type: type } }, + '#withUser':: d.fn(help='User is a SELinux user label that applies to the container.', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { seLinuxOptions+: { user: user } }, + }, + '#withRule':: d.fn(help='rule is the strategy that will dictate the allowable labels that may be set.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { rule: rule }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/supplementalGroupsStrategyOptions.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/supplementalGroupsStrategyOptions.libsonnet new file mode 100644 index 00000000000..f4691e62373 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/policy/v1beta1/supplementalGroupsStrategyOptions.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='supplementalGroupsStrategyOptions', url='', help='SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy.'), + '#withRanges':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.', args=[d.arg(name='ranges', type=d.T.array)]), + withRanges(ranges): { ranges: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRangesMixin':: d.fn(help='ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ranges', type=d.T.array)]), + withRangesMixin(ranges): { ranges+: if std.isArray(v=ranges) then ranges else [ranges] }, + '#withRule':: d.fn(help='rule is the strategy that will dictate what supplemental groups is used in the SecurityContext.', args=[d.arg(name='rule', type=d.T.string)]), + withRule(rule): { rule: rule }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/main.libsonnet new file mode 100644 index 00000000000..f7d3fc1bb5b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='rbac', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1alpha1: (import 'v1alpha1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/aggregationRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/aggregationRule.libsonnet new file mode 100644 index 00000000000..5684fae2f95 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/aggregationRule.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='aggregationRule', url='', help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRole.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRole.libsonnet new file mode 100644 index 00000000000..01d38210b7b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRole.libsonnet @@ -0,0 +1,67 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRole', url='', help='ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding.'), + '#aggregationRule':: d.obj(help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + aggregationRule: { + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrole', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1', + kind: 'ClusterRole', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRoleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRoleBinding.libsonnet new file mode 100644 index 00000000000..86b1808e693 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/clusterRoleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRoleBinding', url='', help='ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1', + kind: 'ClusterRoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/main.libsonnet new file mode 100644 index 00000000000..30ab3dd87c4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/main.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + aggregationRule: (import 'aggregationRule.libsonnet'), + clusterRole: (import 'clusterRole.libsonnet'), + clusterRoleBinding: (import 'clusterRoleBinding.libsonnet'), + policyRule: (import 'policyRule.libsonnet'), + role: (import 'role.libsonnet'), + roleBinding: (import 'roleBinding.libsonnet'), + roleRef: (import 'roleRef.libsonnet'), + subject: (import 'subject.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/policyRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/policyRule.libsonnet new file mode 100644 index 00000000000..0fccf0aa9bd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/policyRule.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='policyRule', url='', help='PolicyRule holds information that describes a policy rule, but does not contain information about who the rule applies to or which namespace the rule applies to.'), + '#withApiGroups':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withNonResourceURLs':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withResourceNames':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNames(resourceNames): { resourceNames: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResourceNamesMixin':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNamesMixin(resourceNames): { resourceNames+: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResources':: d.fn(help='Resources is a list of resources this rule applies to. ResourceAll represents all resources.', args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help='Resources is a list of resources this rule applies to. ResourceAll represents all resources.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/role.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/role.libsonnet new file mode 100644 index 00000000000..e9661d75a97 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/role.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='role', url='', help='Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Role', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1', + kind: 'Role', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this Role', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this Role\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleBinding.libsonnet new file mode 100644 index 00000000000..27e97975dee --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleBinding', url='', help='RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Rolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1', + kind: 'RoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleRef.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleRef.libsonnet new file mode 100644 index 00000000000..ff8c4fae6cd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/roleRef.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleRef', url='', help='RoleRef contains information that points to the role being used'), + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/subject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/subject.libsonnet new file mode 100644 index 00000000000..29265d03d16 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1/subject.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subject', url='', help='Subject contains a reference to the object or user identities a role binding applies to. This can either hold a direct API object reference, or a value for non-objects such as user and group names.'), + '#withApiGroup':: d.fn(help='APIGroup holds the API group of the referenced subject. Defaults to "" for ServiceAccount subjects. Defaults to "rbac.authorization.k8s.io" for User and Group subjects.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind of object being referenced. Values defined by this API group are "User", "Group", and "ServiceAccount". If the Authorizer does not recognized the kind value, the Authorizer should report an error.', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the object being referenced.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace of the referenced object. If the object kind is non-namespace, such as "User" or "Group", and this value is not empty the Authorizer should report an error.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/aggregationRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/aggregationRule.libsonnet new file mode 100644 index 00000000000..5684fae2f95 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/aggregationRule.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='aggregationRule', url='', help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRole.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRole.libsonnet new file mode 100644 index 00000000000..0abdcf931d2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRole.libsonnet @@ -0,0 +1,67 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRole', url='', help='ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22.'), + '#aggregationRule':: d.obj(help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + aggregationRule: { + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrole', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1alpha1', + kind: 'ClusterRole', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRoleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRoleBinding.libsonnet new file mode 100644 index 00000000000..17f3351fc4a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/clusterRoleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRoleBinding', url='', help='ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1alpha1', + kind: 'ClusterRoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..cababc33d00 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/main.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + aggregationRule: (import 'aggregationRule.libsonnet'), + clusterRole: (import 'clusterRole.libsonnet'), + clusterRoleBinding: (import 'clusterRoleBinding.libsonnet'), + policyRule: (import 'policyRule.libsonnet'), + role: (import 'role.libsonnet'), + roleBinding: (import 'roleBinding.libsonnet'), + roleRef: (import 'roleRef.libsonnet'), + subject: (import 'subject.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/policyRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/policyRule.libsonnet new file mode 100644 index 00000000000..0fccf0aa9bd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/policyRule.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='policyRule', url='', help='PolicyRule holds information that describes a policy rule, but does not contain information about who the rule applies to or which namespace the rule applies to.'), + '#withApiGroups':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withNonResourceURLs':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withResourceNames':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNames(resourceNames): { resourceNames: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResourceNamesMixin':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNamesMixin(resourceNames): { resourceNames+: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResources':: d.fn(help='Resources is a list of resources this rule applies to. ResourceAll represents all resources.', args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help='Resources is a list of resources this rule applies to. ResourceAll represents all resources.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/role.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/role.libsonnet new file mode 100644 index 00000000000..3aee62aa96b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/role.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='role', url='', help='Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Role', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1alpha1', + kind: 'Role', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this Role', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this Role\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleBinding.libsonnet new file mode 100644 index 00000000000..1b0fa65f1f2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleBinding', url='', help='RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Rolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1alpha1', + kind: 'RoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleRef.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleRef.libsonnet new file mode 100644 index 00000000000..ff8c4fae6cd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/roleRef.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleRef', url='', help='RoleRef contains information that points to the role being used'), + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/subject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/subject.libsonnet new file mode 100644 index 00000000000..a6e90aeb1ae --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1alpha1/subject.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subject', url='', help='Subject contains a reference to the object or user identities a role binding applies to. This can either hold a direct API object reference, or a value for non-objects such as user and group names.'), + '#withKind':: d.fn(help='Kind of object being referenced. Values defined by this API group are "User", "Group", and "ServiceAccount". If the Authorizer does not recognized the kind value, the Authorizer should report an error.', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the object being referenced.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace of the referenced object. If the object kind is non-namespace, such as "User" or "Group", and this value is not empty the Authorizer should report an error.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/aggregationRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/aggregationRule.libsonnet new file mode 100644 index 00000000000..5684fae2f95 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/aggregationRule.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='aggregationRule', url='', help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRole.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRole.libsonnet new file mode 100644 index 00000000000..acc697644a2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRole.libsonnet @@ -0,0 +1,67 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRole', url='', help='ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRole, and will no longer be served in v1.22.'), + '#aggregationRule':: d.obj(help='AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole'), + aggregationRule: { + '#withClusterRoleSelectors':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectors(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + '#withClusterRoleSelectorsMixin':: d.fn(help="ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='clusterRoleSelectors', type=d.T.array)]), + withClusterRoleSelectorsMixin(clusterRoleSelectors): { aggregationRule+: { clusterRoleSelectors+: if std.isArray(v=clusterRoleSelectors) then clusterRoleSelectors else [clusterRoleSelectors] } }, + }, + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrole', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1beta1', + kind: 'ClusterRole', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this ClusterRole\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRoleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRoleBinding.libsonnet new file mode 100644 index 00000000000..042bd6d9a0b --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/clusterRoleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='clusterRoleBinding', url='', help='ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 ClusterRoleBinding, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Clusterrolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1beta1', + kind: 'ClusterRoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/main.libsonnet new file mode 100644 index 00000000000..67591074ee6 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/main.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + aggregationRule: (import 'aggregationRule.libsonnet'), + clusterRole: (import 'clusterRole.libsonnet'), + clusterRoleBinding: (import 'clusterRoleBinding.libsonnet'), + policyRule: (import 'policyRule.libsonnet'), + role: (import 'role.libsonnet'), + roleBinding: (import 'roleBinding.libsonnet'), + roleRef: (import 'roleRef.libsonnet'), + subject: (import 'subject.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/policyRule.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/policyRule.libsonnet new file mode 100644 index 00000000000..86f9ce8dc1c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/policyRule.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='policyRule', url='', help='PolicyRule holds information that describes a policy rule, but does not contain information about who the rule applies to or which namespace the rule applies to.'), + '#withApiGroups':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroups(apiGroups): { apiGroups: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withApiGroupsMixin':: d.fn(help='APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='apiGroups', type=d.T.array)]), + withApiGroupsMixin(apiGroups): { apiGroups+: if std.isArray(v=apiGroups) then apiGroups else [apiGroups] }, + '#withNonResourceURLs':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLs(nonResourceURLs): { nonResourceURLs: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withNonResourceURLsMixin':: d.fn(help='NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nonResourceURLs', type=d.T.array)]), + withNonResourceURLsMixin(nonResourceURLs): { nonResourceURLs+: if std.isArray(v=nonResourceURLs) then nonResourceURLs else [nonResourceURLs] }, + '#withResourceNames':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNames(resourceNames): { resourceNames: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResourceNamesMixin':: d.fn(help='ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceNames', type=d.T.array)]), + withResourceNamesMixin(resourceNames): { resourceNames+: if std.isArray(v=resourceNames) then resourceNames else [resourceNames] }, + '#withResources':: d.fn(help="Resources is a list of resources this rule applies to. '*' represents all resources in the specified apiGroups. '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.", args=[d.arg(name='resources', type=d.T.array)]), + withResources(resources): { resources: if std.isArray(v=resources) then resources else [resources] }, + '#withResourcesMixin':: d.fn(help="Resources is a list of resources this rule applies to. '*' represents all resources in the specified apiGroups. '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='resources', type=d.T.array)]), + withResourcesMixin(resources): { resources+: if std.isArray(v=resources) then resources else [resources] }, + '#withVerbs':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbs(verbs): { verbs: if std.isArray(v=verbs) then verbs else [verbs] }, + '#withVerbsMixin':: d.fn(help='Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='verbs', type=d.T.array)]), + withVerbsMixin(verbs): { verbs+: if std.isArray(v=verbs) then verbs else [verbs] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/role.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/role.libsonnet new file mode 100644 index 00000000000..37ed05eda7c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/role.libsonnet @@ -0,0 +1,60 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='role', url='', help='Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 Role, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Role', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1beta1', + kind: 'Role', + } + self.metadata.withName(name=name), + '#withRules':: d.fn(help='Rules holds all the PolicyRules for this Role', args=[d.arg(name='rules', type=d.T.array)]), + withRules(rules): { rules: if std.isArray(v=rules) then rules else [rules] }, + '#withRulesMixin':: d.fn(help='Rules holds all the PolicyRules for this Role\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='rules', type=d.T.array)]), + withRulesMixin(rules): { rules+: if std.isArray(v=rules) then rules else [rules] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleBinding.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleBinding.libsonnet new file mode 100644 index 00000000000..fc1fb9b8eb5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleBinding.libsonnet @@ -0,0 +1,69 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleBinding', url='', help='RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. Deprecated in v1.17 in favor of rbac.authorization.k8s.io/v1 RoleBinding, and will no longer be served in v1.22.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Rolebinding', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'rbac.authorization.k8s.io/v1beta1', + kind: 'RoleBinding', + } + self.metadata.withName(name=name), + '#roleRef':: d.obj(help='RoleRef contains information that points to the role being used'), + roleRef: { + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { roleRef+: { apiGroup: apiGroup } }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { roleRef+: { kind: kind } }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { roleRef+: { name: name } }, + }, + '#withSubjects':: d.fn(help='Subjects holds references to the objects the role applies to.', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjects(subjects): { subjects: if std.isArray(v=subjects) then subjects else [subjects] }, + '#withSubjectsMixin':: d.fn(help='Subjects holds references to the objects the role applies to.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='subjects', type=d.T.array)]), + withSubjectsMixin(subjects): { subjects+: if std.isArray(v=subjects) then subjects else [subjects] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleRef.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleRef.libsonnet new file mode 100644 index 00000000000..ff8c4fae6cd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/roleRef.libsonnet @@ -0,0 +1,12 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='roleRef', url='', help='RoleRef contains information that points to the role being used'), + '#withApiGroup':: d.fn(help='APIGroup is the group for the resource being referenced', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind is the type of resource being referenced', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name is the name of resource being referenced', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/subject.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/subject.libsonnet new file mode 100644 index 00000000000..29265d03d16 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/rbac/v1beta1/subject.libsonnet @@ -0,0 +1,14 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='subject', url='', help='Subject contains a reference to the object or user identities a role binding applies to. This can either hold a direct API object reference, or a value for non-objects such as user and group names.'), + '#withApiGroup':: d.fn(help='APIGroup holds the API group of the referenced subject. Defaults to "" for ServiceAccount subjects. Defaults to "rbac.authorization.k8s.io" for User and Group subjects.', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { apiGroup: apiGroup }, + '#withKind':: d.fn(help='Kind of object being referenced. Values defined by this API group are "User", "Group", and "ServiceAccount". If the Authorizer does not recognized the kind value, the Authorizer should report an error.', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { kind: kind }, + '#withName':: d.fn(help='Name of the object being referenced.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='Namespace of the referenced object. If the object kind is non-namespace, such as "User" or "Group", and this value is not empty the Authorizer should report an error.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/main.libsonnet new file mode 100644 index 00000000000..69579538843 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='scheduling', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1alpha1: (import 'v1alpha1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/main.libsonnet new file mode 100644 index 00000000000..76a7ed5aaba --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + priorityClass: (import 'priorityClass.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/priorityClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/priorityClass.libsonnet new file mode 100644 index 00000000000..cbaad8cf0cd --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1/priorityClass.libsonnet @@ -0,0 +1,64 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityClass', url='', help='PriorityClass defines mapping from a priority class name to the priority integer value. The value can be any valid integer.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Priorityclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'scheduling.k8s.io/v1', + kind: 'PriorityClass', + } + self.metadata.withName(name=name), + '#withDescription':: d.fn(help='description is an arbitrary string that usually provides guidelines on when this priority class should be used.', args=[d.arg(name='description', type=d.T.string)]), + withDescription(description): { description: description }, + '#withGlobalDefault':: d.fn(help='globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.', args=[d.arg(name='globalDefault', type=d.T.boolean)]), + withGlobalDefault(globalDefault): { globalDefault: globalDefault }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { preemptionPolicy: preemptionPolicy }, + '#withValue':: d.fn(help='The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.', args=[d.arg(name='value', type=d.T.integer)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..1b042817725 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + priorityClass: (import 'priorityClass.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/priorityClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/priorityClass.libsonnet new file mode 100644 index 00000000000..83a38c4fa21 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1alpha1/priorityClass.libsonnet @@ -0,0 +1,64 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityClass', url='', help='DEPRECATED - This group version of PriorityClass is deprecated by scheduling.k8s.io/v1/PriorityClass. PriorityClass defines mapping from a priority class name to the priority integer value. The value can be any valid integer.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Priorityclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'scheduling.k8s.io/v1alpha1', + kind: 'PriorityClass', + } + self.metadata.withName(name=name), + '#withDescription':: d.fn(help='description is an arbitrary string that usually provides guidelines on when this priority class should be used.', args=[d.arg(name='description', type=d.T.string)]), + withDescription(description): { description: description }, + '#withGlobalDefault':: d.fn(help='globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.', args=[d.arg(name='globalDefault', type=d.T.boolean)]), + withGlobalDefault(globalDefault): { globalDefault: globalDefault }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { preemptionPolicy: preemptionPolicy }, + '#withValue':: d.fn(help='The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.', args=[d.arg(name='value', type=d.T.integer)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/main.libsonnet new file mode 100644 index 00000000000..3b681adec03 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + priorityClass: (import 'priorityClass.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/priorityClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/priorityClass.libsonnet new file mode 100644 index 00000000000..d9002396a62 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/scheduling/v1beta1/priorityClass.libsonnet @@ -0,0 +1,64 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='priorityClass', url='', help='DEPRECATED - This group version of PriorityClass is deprecated by scheduling.k8s.io/v1/PriorityClass. PriorityClass defines mapping from a priority class name to the priority integer value. The value can be any valid integer.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Priorityclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'scheduling.k8s.io/v1beta1', + kind: 'PriorityClass', + } + self.metadata.withName(name=name), + '#withDescription':: d.fn(help='description is an arbitrary string that usually provides guidelines on when this priority class should be used.', args=[d.arg(name='description', type=d.T.string)]), + withDescription(description): { description: description }, + '#withGlobalDefault':: d.fn(help='globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority.', args=[d.arg(name='globalDefault', type=d.T.boolean)]), + withGlobalDefault(globalDefault): { globalDefault: globalDefault }, + '#withPreemptionPolicy':: d.fn(help='PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset. This field is beta-level, gated by the NonPreemptingPriority feature-gate.', args=[d.arg(name='preemptionPolicy', type=d.T.string)]), + withPreemptionPolicy(preemptionPolicy): { preemptionPolicy: preemptionPolicy }, + '#withValue':: d.fn(help='The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec.', args=[d.arg(name='value', type=d.T.integer)]), + withValue(value): { value: value }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/main.libsonnet new file mode 100644 index 00000000000..c65474321d0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/main.libsonnet @@ -0,0 +1,5 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='settings', url='', help=''), + v1alpha1: (import 'v1alpha1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..8e9ca8de809 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + podPreset: (import 'podPreset.libsonnet'), + podPresetSpec: (import 'podPresetSpec.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPreset.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPreset.libsonnet new file mode 100644 index 00000000000..72eda7f3e51 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPreset.libsonnet @@ -0,0 +1,86 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podPreset', url='', help='PodPreset is a policy resource that defines additional runtime requirements for a Pod.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Podpreset', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'settings.k8s.io/v1alpha1', + kind: 'PodPreset', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='PodPresetSpec is a description of a pod preset.'), + spec: { + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withEnv':: d.fn(help='Env defines the collection of EnvVar to inject into containers.', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } }, + '#withEnvFrom':: d.fn(help='EnvFrom defines the collection of EnvFromSource to inject into containers.', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { spec+: { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] } }, + '#withEnvFromMixin':: d.fn(help='EnvFrom defines the collection of EnvFromSource to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { spec+: { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] } }, + '#withEnvMixin':: d.fn(help='Env defines the collection of EnvVar to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { spec+: { env+: if std.isArray(v=env) then env else [env] } }, + '#withVolumeMounts':: d.fn(help='VolumeMounts defines the collection of VolumeMount to inject into containers.', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { spec+: { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } }, + '#withVolumeMountsMixin':: d.fn(help='VolumeMounts defines the collection of VolumeMount to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { spec+: { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } }, + '#withVolumes':: d.fn(help='Volumes defines the collection of Volume to inject into the pod.', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } }, + '#withVolumesMixin':: d.fn(help='Volumes defines the collection of Volume to inject into the pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPresetSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPresetSpec.libsonnet new file mode 100644 index 00000000000..0fc8c5c651f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/settings/v1alpha1/podPresetSpec.libsonnet @@ -0,0 +1,33 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podPresetSpec', url='', help='PodPresetSpec is a description of a pod preset.'), + '#selector':: d.obj(help='A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects.'), + selector: { + '#withMatchExpressions':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='matchExpressions is a list of label selector requirements. The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { selector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { selector+: { matchLabels+: matchLabels } }, + }, + '#withEnv':: d.fn(help='Env defines the collection of EnvVar to inject into containers.', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='EnvFrom defines the collection of EnvFromSource to inject into containers.', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='EnvFrom defines the collection of EnvFromSource to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='Env defines the collection of EnvVar to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withVolumeMounts':: d.fn(help='VolumeMounts defines the collection of VolumeMount to inject into containers.', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help='VolumeMounts defines the collection of VolumeMount to inject into containers.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumes':: d.fn(help='Volumes defines the collection of Volume to inject into the pod.', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { volumes: if std.isArray(v=volumes) then volumes else [volumes] }, + '#withVolumesMixin':: d.fn(help='Volumes defines the collection of Volume to inject into the pod.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { volumes+: if std.isArray(v=volumes) then volumes else [volumes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/main.libsonnet new file mode 100644 index 00000000000..5b2186168a5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/main.libsonnet @@ -0,0 +1,7 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='storage', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1alpha1: (import 'v1alpha1/main.libsonnet'), + v1beta1: (import 'v1beta1/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriver.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriver.libsonnet new file mode 100644 index 00000000000..0cda0b53715 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriver.libsonnet @@ -0,0 +1,71 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiDriver', url='', help='CSIDriver captures information about a Container Storage Interface (CSI) volume driver deployed on the cluster. Kubernetes attach detach controller uses this object to determine whether attach is required. Kubelet uses this object to determine whether pod information needs to be passed on mount. CSIDriver objects are non-namespaced.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Csidriver', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1', + kind: 'CsiDriver', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CSIDriverSpec is the specification of a CSIDriver.'), + spec: { + '#withAttachRequired':: d.fn(help='attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.', args=[d.arg(name='attachRequired', type=d.T.boolean)]), + withAttachRequired(attachRequired): { spec+: { attachRequired: attachRequired } }, + '#withFsGroupPolicy':: d.fn(help='Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.', args=[d.arg(name='fsGroupPolicy', type=d.T.string)]), + withFsGroupPolicy(fsGroupPolicy): { spec+: { fsGroupPolicy: fsGroupPolicy } }, + '#withPodInfoOnMount':: d.fn(help="If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. 'csi.storage.k8s.io/pod.name': pod.Name 'csi.storage.k8s.io/pod.namespace': pod.Namespace 'csi.storage.k8s.io/pod.uid': string(pod.UID) 'csi.storage.k8s.io/ephemeral': 'true' iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise 'false'\n\n'csi.storage.k8s.io/ephemeral' is a new feature in Kubernetes 1.16. It is only required for drivers which support both the 'Persistent' and 'Ephemeral' VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", args=[d.arg(name='podInfoOnMount', type=d.T.boolean)]), + withPodInfoOnMount(podInfoOnMount): { spec+: { podInfoOnMount: podInfoOnMount } }, + '#withStorageCapacity':: d.fn(help='If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.', args=[d.arg(name='storageCapacity', type=d.T.boolean)]), + withStorageCapacity(storageCapacity): { spec+: { storageCapacity: storageCapacity } }, + '#withVolumeLifecycleModes':: d.fn(help='volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future. This field is beta.', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModes(volumeLifecycleModes): { spec+: { volumeLifecycleModes: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] } }, + '#withVolumeLifecycleModesMixin':: d.fn(help='volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future. This field is beta.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModesMixin(volumeLifecycleModes): { spec+: { volumeLifecycleModes+: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriverSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriverSpec.libsonnet new file mode 100644 index 00000000000..86f14f4b12f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiDriverSpec.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiDriverSpec', url='', help='CSIDriverSpec is the specification of a CSIDriver.'), + '#withAttachRequired':: d.fn(help='attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.', args=[d.arg(name='attachRequired', type=d.T.boolean)]), + withAttachRequired(attachRequired): { attachRequired: attachRequired }, + '#withFsGroupPolicy':: d.fn(help='Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.', args=[d.arg(name='fsGroupPolicy', type=d.T.string)]), + withFsGroupPolicy(fsGroupPolicy): { fsGroupPolicy: fsGroupPolicy }, + '#withPodInfoOnMount':: d.fn(help="If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. 'csi.storage.k8s.io/pod.name': pod.Name 'csi.storage.k8s.io/pod.namespace': pod.Namespace 'csi.storage.k8s.io/pod.uid': string(pod.UID) 'csi.storage.k8s.io/ephemeral': 'true' iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise 'false'\n\n'csi.storage.k8s.io/ephemeral' is a new feature in Kubernetes 1.16. It is only required for drivers which support both the 'Persistent' and 'Ephemeral' VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", args=[d.arg(name='podInfoOnMount', type=d.T.boolean)]), + withPodInfoOnMount(podInfoOnMount): { podInfoOnMount: podInfoOnMount }, + '#withStorageCapacity':: d.fn(help='If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.', args=[d.arg(name='storageCapacity', type=d.T.boolean)]), + withStorageCapacity(storageCapacity): { storageCapacity: storageCapacity }, + '#withVolumeLifecycleModes':: d.fn(help='volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future. This field is beta.', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModes(volumeLifecycleModes): { volumeLifecycleModes: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] }, + '#withVolumeLifecycleModesMixin':: d.fn(help='volumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future. This field is beta.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModesMixin(volumeLifecycleModes): { volumeLifecycleModes+: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNode.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNode.libsonnet new file mode 100644 index 00000000000..d2e44067653 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNode.libsonnet @@ -0,0 +1,63 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNode', url='', help="CSINode holds information about all CSI drivers installed on a node. CSI drivers do not need to create the CSINode object directly. As long as they use the node-driver-registrar sidecar container, the kubelet will automatically populate the CSINode object for the CSI driver as part of kubelet plugin registration. CSINode has the same name as a node. If the object is missing, it means either there are no CSI Drivers available on the node, or the Kubelet version is low enough that it doesn't create this object. CSINode has an OwnerReference that points to the corresponding node object."), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Csinode', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1', + kind: 'CsiNode', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CSINodeSpec holds information about the specification of all CSI drivers installed on a node'), + spec: { + '#withDrivers':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.', args=[d.arg(name='drivers', type=d.T.array)]), + withDrivers(drivers): { spec+: { drivers: if std.isArray(v=drivers) then drivers else [drivers] } }, + '#withDriversMixin':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drivers', type=d.T.array)]), + withDriversMixin(drivers): { spec+: { drivers+: if std.isArray(v=drivers) then drivers else [drivers] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeDriver.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeDriver.libsonnet new file mode 100644 index 00000000000..3acd27eaba4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeDriver.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNodeDriver', url='', help='CSINodeDriver holds information about the specification of one CSI driver installed on a node'), + '#allocatable':: d.obj(help='VolumeNodeResources is a set of resource limits for scheduling of volumes.'), + allocatable: { + '#withCount':: d.fn(help='Maximum number of unique volumes managed by the CSI driver that can be used on a node. A volume that is both attached and mounted on a node is considered to be used once, not twice. The same rule applies for a unique volume that is shared among multiple pods on the same node. If this field is not specified, then the supported number of volumes on this node is unbounded.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { allocatable+: { count: count } }, + }, + '#withName':: d.fn(help='This is the name of the CSI driver that this object refers to. This MUST be the same name returned by the CSI GetPluginName() call for that driver.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNodeID':: d.fn(help='nodeID of the node from the driver point of view. This field enables Kubernetes to communicate with storage systems that do not share the same nomenclature for nodes. For example, Kubernetes may refer to a given node as "node1", but the storage system may refer to the same node as "nodeA". When Kubernetes issues a command to the storage system to attach a volume to a specific node, it can use this field to refer to the node name using the ID that the storage system will understand, e.g. "nodeA" instead of "node1". This field is required.', args=[d.arg(name='nodeID', type=d.T.string)]), + withNodeID(nodeID): { nodeID: nodeID }, + '#withTopologyKeys':: d.fn(help='topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology.', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeys(topologyKeys): { topologyKeys: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#withTopologyKeysMixin':: d.fn(help='topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeysMixin(topologyKeys): { topologyKeys+: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeSpec.libsonnet new file mode 100644 index 00000000000..c38b1d79169 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/csiNodeSpec.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNodeSpec', url='', help='CSINodeSpec holds information about the specification of all CSI drivers installed on a node'), + '#withDrivers':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.', args=[d.arg(name='drivers', type=d.T.array)]), + withDrivers(drivers): { drivers: if std.isArray(v=drivers) then drivers else [drivers] }, + '#withDriversMixin':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drivers', type=d.T.array)]), + withDriversMixin(drivers): { drivers+: if std.isArray(v=drivers) then drivers else [drivers] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/main.libsonnet new file mode 100644 index 00000000000..c45e34d497c --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/main.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + csiDriver: (import 'csiDriver.libsonnet'), + csiDriverSpec: (import 'csiDriverSpec.libsonnet'), + csiNode: (import 'csiNode.libsonnet'), + csiNodeDriver: (import 'csiNodeDriver.libsonnet'), + csiNodeSpec: (import 'csiNodeSpec.libsonnet'), + storageClass: (import 'storageClass.libsonnet'), + volumeAttachment: (import 'volumeAttachment.libsonnet'), + volumeAttachmentSource: (import 'volumeAttachmentSource.libsonnet'), + volumeAttachmentSpec: (import 'volumeAttachmentSpec.libsonnet'), + volumeAttachmentStatus: (import 'volumeAttachmentStatus.libsonnet'), + volumeError: (import 'volumeError.libsonnet'), + volumeNodeResources: (import 'volumeNodeResources.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/storageClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/storageClass.libsonnet new file mode 100644 index 00000000000..4a5e651ac44 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/storageClass.libsonnet @@ -0,0 +1,76 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='storageClass', url='', help='StorageClass describes the parameters for a class of storage for which PersistentVolumes can be dynamically provisioned.\n\nStorageClasses are non-namespaced; the name of the storage class according to etcd is in ObjectMeta.Name.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Storageclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1', + kind: 'StorageClass', + } + self.metadata.withName(name=name), + '#withAllowVolumeExpansion':: d.fn(help='AllowVolumeExpansion shows whether the storage class allow volume expand', args=[d.arg(name='allowVolumeExpansion', type=d.T.boolean)]), + withAllowVolumeExpansion(allowVolumeExpansion): { allowVolumeExpansion: allowVolumeExpansion }, + '#withAllowedTopologies':: d.fn(help='Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature.', args=[d.arg(name='allowedTopologies', type=d.T.array)]), + withAllowedTopologies(allowedTopologies): { allowedTopologies: if std.isArray(v=allowedTopologies) then allowedTopologies else [allowedTopologies] }, + '#withAllowedTopologiesMixin':: d.fn(help='Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedTopologies', type=d.T.array)]), + withAllowedTopologiesMixin(allowedTopologies): { allowedTopologies+: if std.isArray(v=allowedTopologies) then allowedTopologies else [allowedTopologies] }, + '#withMountOptions':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid.', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withMountOptionsMixin':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withParameters':: d.fn(help='Parameters holds the parameters for the provisioner that should create volumes of this storage class.', args=[d.arg(name='parameters', type=d.T.object)]), + withParameters(parameters): { parameters: parameters }, + '#withParametersMixin':: d.fn(help='Parameters holds the parameters for the provisioner that should create volumes of this storage class.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='parameters', type=d.T.object)]), + withParametersMixin(parameters): { parameters+: parameters }, + '#withProvisioner':: d.fn(help='Provisioner indicates the type of the provisioner.', args=[d.arg(name='provisioner', type=d.T.string)]), + withProvisioner(provisioner): { provisioner: provisioner }, + '#withReclaimPolicy':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with this reclaimPolicy. Defaults to Delete.', args=[d.arg(name='reclaimPolicy', type=d.T.string)]), + withReclaimPolicy(reclaimPolicy): { reclaimPolicy: reclaimPolicy }, + '#withVolumeBindingMode':: d.fn(help='VolumeBindingMode indicates how PersistentVolumeClaims should be provisioned and bound. When unset, VolumeBindingImmediate is used. This field is only honored by servers that enable the VolumeScheduling feature.', args=[d.arg(name='volumeBindingMode', type=d.T.string)]), + withVolumeBindingMode(volumeBindingMode): { volumeBindingMode: volumeBindingMode }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachment.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachment.libsonnet new file mode 100644 index 00000000000..6aff6fdecf0 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachment.libsonnet @@ -0,0 +1,479 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachment', url='', help='VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node.\n\nVolumeAttachment objects are non-namespaced.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Volumeattachment', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1', + kind: 'VolumeAttachment', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + spec: { + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity: capacity } } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity+: capacity } } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { spec+: { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { spec+: { source+: { persistentVolumeName: persistentVolumeName } } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { spec+: { attacher: attacher } }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { nodeName: nodeName } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSource.libsonnet new file mode 100644 index 00000000000..65b29e14e10 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSource.libsonnet @@ -0,0 +1,419 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSource', url='', help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { azureDisk+: { kind: kind } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { cephfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { cephfs+: { user: user } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { cinder+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { claimRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { claimRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { claimRef+: { uid: uid } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { csi+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { csi+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { fc+: { fsType: fsType } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { fc+: { lun: lun } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { flexVolume+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { inlineVolumeSpec+: { flexVolume+: { options: options } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { inlineVolumeSpec+: { flexVolume+: { options+: options } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { glusterfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { hostPath+: { path: path } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { inlineVolumeSpec+: { hostPath+: { type: type } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { iscsi+: { lun: lun } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { 'local'+: { path: path } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { nfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { inlineVolumeSpec+: { nfs+: { server: server } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { inlineVolumeSpec+: { quobyte+: { group: group } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { inlineVolumeSpec+: { quobyte+: { registry: registry } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { quobyte+: { user: user } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { inlineVolumeSpec+: { quobyte+: { volume: volume } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { rbd+: { fsType: fsType } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { inlineVolumeSpec+: { rbd+: { image: image } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { inlineVolumeSpec+: { rbd+: { keyring: keyring } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { inlineVolumeSpec+: { rbd+: { pool: pool } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { rbd+: { user: user } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { inlineVolumeSpec+: { scaleIO+: { system: system } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { storageos+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { inlineVolumeSpec+: { capacity: capacity } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { inlineVolumeSpec+: { capacity+: capacity } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { inlineVolumeSpec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { inlineVolumeSpec+: { volumeMode: volumeMode } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { persistentVolumeName: persistentVolumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSpec.libsonnet new file mode 100644 index 00000000000..29f116537cf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentSpec.libsonnet @@ -0,0 +1,426 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSpec', url='', help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { source+: { inlineVolumeSpec+: { capacity: capacity } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { source+: { inlineVolumeSpec+: { capacity+: capacity } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { source+: { persistentVolumeName: persistentVolumeName } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { attacher: attacher }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { nodeName: nodeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentStatus.libsonnet new file mode 100644 index 00000000000..7d14c51ec30 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeAttachmentStatus.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentStatus', url='', help='VolumeAttachmentStatus is the status of a VolumeAttachment request.'), + '#attachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + attachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { attachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { attachError+: { time: time } }, + }, + '#detachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + detachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { detachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { detachError+: { time: time } }, + }, + '#withAttached':: d.fn(help='Indicates the volume is successfully attached. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attached', type=d.T.boolean)]), + withAttached(attached): { attached: attached }, + '#withAttachmentMetadata':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadata(attachmentMetadata): { attachmentMetadata: attachmentMetadata }, + '#withAttachmentMetadataMixin':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadataMixin(attachmentMetadata): { attachmentMetadata+: attachmentMetadata }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeError.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeError.libsonnet new file mode 100644 index 00000000000..aea460dc2e1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeError.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeError', url='', help='VolumeError captures an error encountered during a volume operation.'), + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { time: time }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeNodeResources.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeNodeResources.libsonnet new file mode 100644 index 00000000000..3406637ef9e --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1/volumeNodeResources.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeNodeResources', url='', help='VolumeNodeResources is a set of resource limits for scheduling of volumes.'), + '#withCount':: d.fn(help='Maximum number of unique volumes managed by the CSI driver that can be used on a node. A volume that is both attached and mounted on a node is considered to be used once, not twice. The same rule applies for a unique volume that is shared among multiple pods on the same node. If this field is not specified, then the supported number of volumes on this node is unbounded.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/main.libsonnet new file mode 100644 index 00000000000..04b1c24b2b4 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/main.libsonnet @@ -0,0 +1,9 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1alpha1', url='', help=''), + volumeAttachment: (import 'volumeAttachment.libsonnet'), + volumeAttachmentSource: (import 'volumeAttachmentSource.libsonnet'), + volumeAttachmentSpec: (import 'volumeAttachmentSpec.libsonnet'), + volumeAttachmentStatus: (import 'volumeAttachmentStatus.libsonnet'), + volumeError: (import 'volumeError.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachment.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachment.libsonnet new file mode 100644 index 00000000000..3a251d3f536 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachment.libsonnet @@ -0,0 +1,479 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachment', url='', help='VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node.\n\nVolumeAttachment objects are non-namespaced.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Volumeattachment', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1alpha1', + kind: 'VolumeAttachment', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + spec: { + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity: capacity } } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity+: capacity } } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { spec+: { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { spec+: { source+: { persistentVolumeName: persistentVolumeName } } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { spec+: { attacher: attacher } }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { nodeName: nodeName } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSource.libsonnet new file mode 100644 index 00000000000..65b29e14e10 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSource.libsonnet @@ -0,0 +1,419 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSource', url='', help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { azureDisk+: { kind: kind } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { cephfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { cephfs+: { user: user } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { cinder+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { claimRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { claimRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { claimRef+: { uid: uid } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { csi+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { csi+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { fc+: { fsType: fsType } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { fc+: { lun: lun } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { flexVolume+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { inlineVolumeSpec+: { flexVolume+: { options: options } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { inlineVolumeSpec+: { flexVolume+: { options+: options } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { glusterfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { hostPath+: { path: path } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { inlineVolumeSpec+: { hostPath+: { type: type } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { iscsi+: { lun: lun } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { 'local'+: { path: path } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { nfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { inlineVolumeSpec+: { nfs+: { server: server } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { inlineVolumeSpec+: { quobyte+: { group: group } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { inlineVolumeSpec+: { quobyte+: { registry: registry } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { quobyte+: { user: user } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { inlineVolumeSpec+: { quobyte+: { volume: volume } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { rbd+: { fsType: fsType } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { inlineVolumeSpec+: { rbd+: { image: image } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { inlineVolumeSpec+: { rbd+: { keyring: keyring } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { inlineVolumeSpec+: { rbd+: { pool: pool } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { rbd+: { user: user } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { inlineVolumeSpec+: { scaleIO+: { system: system } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { storageos+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { inlineVolumeSpec+: { capacity: capacity } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { inlineVolumeSpec+: { capacity+: capacity } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { inlineVolumeSpec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { inlineVolumeSpec+: { volumeMode: volumeMode } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { persistentVolumeName: persistentVolumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSpec.libsonnet new file mode 100644 index 00000000000..29f116537cf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentSpec.libsonnet @@ -0,0 +1,426 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSpec', url='', help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { source+: { inlineVolumeSpec+: { capacity: capacity } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { source+: { inlineVolumeSpec+: { capacity+: capacity } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { source+: { persistentVolumeName: persistentVolumeName } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { attacher: attacher }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { nodeName: nodeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentStatus.libsonnet new file mode 100644 index 00000000000..9b56054a42f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeAttachmentStatus.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentStatus', url='', help='VolumeAttachmentStatus is the status of a VolumeAttachment request.'), + '#attachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + attachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string maybe logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { attachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { attachError+: { time: time } }, + }, + '#detachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + detachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string maybe logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { detachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { detachError+: { time: time } }, + }, + '#withAttached':: d.fn(help='Indicates the volume is successfully attached. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attached', type=d.T.boolean)]), + withAttached(attached): { attached: attached }, + '#withAttachmentMetadata':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadata(attachmentMetadata): { attachmentMetadata: attachmentMetadata }, + '#withAttachmentMetadataMixin':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadataMixin(attachmentMetadata): { attachmentMetadata+: attachmentMetadata }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeError.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeError.libsonnet new file mode 100644 index 00000000000..d0aa7b4405f --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1alpha1/volumeError.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeError', url='', help='VolumeError captures an error encountered during a volume operation.'), + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string maybe logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { time: time }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriver.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriver.libsonnet new file mode 100644 index 00000000000..2c77fbd2c57 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriver.libsonnet @@ -0,0 +1,71 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiDriver', url='', help='CSIDriver captures information about a Container Storage Interface (CSI) volume driver deployed on the cluster. CSI drivers do not need to create the CSIDriver object directly. Instead they may use the cluster-driver-registrar sidecar container. When deployed with a CSI driver it automatically creates a CSIDriver object representing the driver. Kubernetes attach detach controller uses this object to determine whether attach is required. Kubelet uses this object to determine whether pod information needs to be passed on mount. CSIDriver objects are non-namespaced.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Csidriver', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1beta1', + kind: 'CsiDriver', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CSIDriverSpec is the specification of a CSIDriver.'), + spec: { + '#withAttachRequired':: d.fn(help='attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.', args=[d.arg(name='attachRequired', type=d.T.boolean)]), + withAttachRequired(attachRequired): { spec+: { attachRequired: attachRequired } }, + '#withFsGroupPolicy':: d.fn(help='Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.', args=[d.arg(name='fsGroupPolicy', type=d.T.string)]), + withFsGroupPolicy(fsGroupPolicy): { spec+: { fsGroupPolicy: fsGroupPolicy } }, + '#withPodInfoOnMount':: d.fn(help="If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. 'csi.storage.k8s.io/pod.name': pod.Name 'csi.storage.k8s.io/pod.namespace': pod.Namespace 'csi.storage.k8s.io/pod.uid': string(pod.UID) 'csi.storage.k8s.io/ephemeral': 'true' iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise 'false'\n\n'csi.storage.k8s.io/ephemeral' is a new feature in Kubernetes 1.16. It is only required for drivers which support both the 'Persistent' and 'Ephemeral' VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", args=[d.arg(name='podInfoOnMount', type=d.T.boolean)]), + withPodInfoOnMount(podInfoOnMount): { spec+: { podInfoOnMount: podInfoOnMount } }, + '#withStorageCapacity':: d.fn(help='If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.', args=[d.arg(name='storageCapacity', type=d.T.boolean)]), + withStorageCapacity(storageCapacity): { spec+: { storageCapacity: storageCapacity } }, + '#withVolumeLifecycleModes':: d.fn(help='VolumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future.', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModes(volumeLifecycleModes): { spec+: { volumeLifecycleModes: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] } }, + '#withVolumeLifecycleModesMixin':: d.fn(help='VolumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModesMixin(volumeLifecycleModes): { spec+: { volumeLifecycleModes+: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriverSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriverSpec.libsonnet new file mode 100644 index 00000000000..0ec18c50ef2 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiDriverSpec.libsonnet @@ -0,0 +1,18 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiDriverSpec', url='', help='CSIDriverSpec is the specification of a CSIDriver.'), + '#withAttachRequired':: d.fn(help='attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called.', args=[d.arg(name='attachRequired', type=d.T.boolean)]), + withAttachRequired(attachRequired): { attachRequired: attachRequired }, + '#withFsGroupPolicy':: d.fn(help='Defines if the underlying volume supports changing ownership and permission of the volume before being mounted. Refer to the specific FSGroupPolicy values for additional details. This field is alpha-level, and is only honored by servers that enable the CSIVolumeFSGroupPolicy feature gate.', args=[d.arg(name='fsGroupPolicy', type=d.T.string)]), + withFsGroupPolicy(fsGroupPolicy): { fsGroupPolicy: fsGroupPolicy }, + '#withPodInfoOnMount':: d.fn(help="If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. 'csi.storage.k8s.io/pod.name': pod.Name 'csi.storage.k8s.io/pod.namespace': pod.Namespace 'csi.storage.k8s.io/pod.uid': string(pod.UID) 'csi.storage.k8s.io/ephemeral': 'true' iff the volume is an ephemeral inline volume\n defined by a CSIVolumeSource, otherwise 'false'\n\n'csi.storage.k8s.io/ephemeral' is a new feature in Kubernetes 1.16. It is only required for drivers which support both the 'Persistent' and 'Ephemeral' VolumeLifecycleMode. Other drivers can leave pod info disabled and/or ignore this field. As Kubernetes 1.15 doesn't support this field, drivers can only support one mode when deployed on such a cluster and the deployment determines which mode that is, for example via a command line parameter of the driver.", args=[d.arg(name='podInfoOnMount', type=d.T.boolean)]), + withPodInfoOnMount(podInfoOnMount): { podInfoOnMount: podInfoOnMount }, + '#withStorageCapacity':: d.fn(help='If set to true, storageCapacity indicates that the CSI volume driver wants pod scheduling to consider the storage capacity that the driver deployment will report by creating CSIStorageCapacity objects with capacity information.\n\nThe check can be enabled immediately when deploying a driver. In that case, provisioning new volumes with late binding will pause until the driver deployment has published some suitable CSIStorageCapacity object.\n\nAlternatively, the driver can be deployed with the field unset or false and it can be flipped later when storage capacity information has been published.\n\nThis is an alpha field and only available when the CSIStorageCapacity feature is enabled. The default is false.', args=[d.arg(name='storageCapacity', type=d.T.boolean)]), + withStorageCapacity(storageCapacity): { storageCapacity: storageCapacity }, + '#withVolumeLifecycleModes':: d.fn(help='VolumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future.', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModes(volumeLifecycleModes): { volumeLifecycleModes: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] }, + '#withVolumeLifecycleModesMixin':: d.fn(help='VolumeLifecycleModes defines what kind of volumes this CSI volume driver supports. The default if the list is empty is "Persistent", which is the usage defined by the CSI specification and implemented in Kubernetes via the usual PV/PVC mechanism. The other mode is "Ephemeral". In this mode, volumes are defined inline inside the pod spec with CSIVolumeSource and their lifecycle is tied to the lifecycle of that pod. A driver has to be aware of this because it is only going to get a NodePublishVolume call for such a volume. For more information about implementing this mode, see https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html A driver can support one or more of these modes and more modes may be added in the future.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeLifecycleModes', type=d.T.array)]), + withVolumeLifecycleModesMixin(volumeLifecycleModes): { volumeLifecycleModes+: if std.isArray(v=volumeLifecycleModes) then volumeLifecycleModes else [volumeLifecycleModes] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNode.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNode.libsonnet new file mode 100644 index 00000000000..e14e33ee286 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNode.libsonnet @@ -0,0 +1,63 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNode', url='', help="DEPRECATED - This group version of CSINode is deprecated by storage/v1/CSINode. See the release notes for more information. CSINode holds information about all CSI drivers installed on a node. CSI drivers do not need to create the CSINode object directly. As long as they use the node-driver-registrar sidecar container, the kubelet will automatically populate the CSINode object for the CSI driver as part of kubelet plugin registration. CSINode has the same name as a node. If the object is missing, it means either there are no CSI Drivers available on the node, or the Kubelet version is low enough that it doesn't create this object. CSINode has an OwnerReference that points to the corresponding node object."), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Csinode', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1beta1', + kind: 'CsiNode', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='CSINodeSpec holds information about the specification of all CSI drivers installed on a node'), + spec: { + '#withDrivers':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.', args=[d.arg(name='drivers', type=d.T.array)]), + withDrivers(drivers): { spec+: { drivers: if std.isArray(v=drivers) then drivers else [drivers] } }, + '#withDriversMixin':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drivers', type=d.T.array)]), + withDriversMixin(drivers): { spec+: { drivers+: if std.isArray(v=drivers) then drivers else [drivers] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeDriver.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeDriver.libsonnet new file mode 100644 index 00000000000..b5ed5743360 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeDriver.libsonnet @@ -0,0 +1,19 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNodeDriver', url='', help='CSINodeDriver holds information about the specification of one CSI driver installed on a node'), + '#allocatable':: d.obj(help='VolumeNodeResources is a set of resource limits for scheduling of volumes.'), + allocatable: { + '#withCount':: d.fn(help='Maximum number of unique volumes managed by the CSI driver that can be used on a node. A volume that is both attached and mounted on a node is considered to be used once, not twice. The same rule applies for a unique volume that is shared among multiple pods on the same node. If this field is nil, then the supported number of volumes on this node is unbounded.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { allocatable+: { count: count } }, + }, + '#withName':: d.fn(help='This is the name of the CSI driver that this object refers to. This MUST be the same name returned by the CSI GetPluginName() call for that driver.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNodeID':: d.fn(help='nodeID of the node from the driver point of view. This field enables Kubernetes to communicate with storage systems that do not share the same nomenclature for nodes. For example, Kubernetes may refer to a given node as "node1", but the storage system may refer to the same node as "nodeA". When Kubernetes issues a command to the storage system to attach a volume to a specific node, it can use this field to refer to the node name using the ID that the storage system will understand, e.g. "nodeA" instead of "node1". This field is required.', args=[d.arg(name='nodeID', type=d.T.string)]), + withNodeID(nodeID): { nodeID: nodeID }, + '#withTopologyKeys':: d.fn(help='topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology.', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeys(topologyKeys): { topologyKeys: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#withTopologyKeysMixin':: d.fn(help='topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologyKeys', type=d.T.array)]), + withTopologyKeysMixin(topologyKeys): { topologyKeys+: if std.isArray(v=topologyKeys) then topologyKeys else [topologyKeys] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeSpec.libsonnet new file mode 100644 index 00000000000..c38b1d79169 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/csiNodeSpec.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='csiNodeSpec', url='', help='CSINodeSpec holds information about the specification of all CSI drivers installed on a node'), + '#withDrivers':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.', args=[d.arg(name='drivers', type=d.T.array)]), + withDrivers(drivers): { drivers: if std.isArray(v=drivers) then drivers else [drivers] }, + '#withDriversMixin':: d.fn(help='drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drivers', type=d.T.array)]), + withDriversMixin(drivers): { drivers+: if std.isArray(v=drivers) then drivers else [drivers] }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/main.libsonnet new file mode 100644 index 00000000000..01f4a63f8bf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/main.libsonnet @@ -0,0 +1,16 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1beta1', url='', help=''), + csiDriver: (import 'csiDriver.libsonnet'), + csiDriverSpec: (import 'csiDriverSpec.libsonnet'), + csiNode: (import 'csiNode.libsonnet'), + csiNodeDriver: (import 'csiNodeDriver.libsonnet'), + csiNodeSpec: (import 'csiNodeSpec.libsonnet'), + storageClass: (import 'storageClass.libsonnet'), + volumeAttachment: (import 'volumeAttachment.libsonnet'), + volumeAttachmentSource: (import 'volumeAttachmentSource.libsonnet'), + volumeAttachmentSpec: (import 'volumeAttachmentSpec.libsonnet'), + volumeAttachmentStatus: (import 'volumeAttachmentStatus.libsonnet'), + volumeError: (import 'volumeError.libsonnet'), + volumeNodeResources: (import 'volumeNodeResources.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/storageClass.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/storageClass.libsonnet new file mode 100644 index 00000000000..2e74cab5505 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/storageClass.libsonnet @@ -0,0 +1,76 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='storageClass', url='', help='StorageClass describes the parameters for a class of storage for which PersistentVolumes can be dynamically provisioned.\n\nStorageClasses are non-namespaced; the name of the storage class according to etcd is in ObjectMeta.Name.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Storageclass', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1beta1', + kind: 'StorageClass', + } + self.metadata.withName(name=name), + '#withAllowVolumeExpansion':: d.fn(help='AllowVolumeExpansion shows whether the storage class allow volume expand', args=[d.arg(name='allowVolumeExpansion', type=d.T.boolean)]), + withAllowVolumeExpansion(allowVolumeExpansion): { allowVolumeExpansion: allowVolumeExpansion }, + '#withAllowedTopologies':: d.fn(help='Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature.', args=[d.arg(name='allowedTopologies', type=d.T.array)]), + withAllowedTopologies(allowedTopologies): { allowedTopologies: if std.isArray(v=allowedTopologies) then allowedTopologies else [allowedTopologies] }, + '#withAllowedTopologiesMixin':: d.fn(help='Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='allowedTopologies', type=d.T.array)]), + withAllowedTopologiesMixin(allowedTopologies): { allowedTopologies+: if std.isArray(v=allowedTopologies) then allowedTopologies else [allowedTopologies] }, + '#withMountOptions':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid.', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withMountOptionsMixin':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] }, + '#withParameters':: d.fn(help='Parameters holds the parameters for the provisioner that should create volumes of this storage class.', args=[d.arg(name='parameters', type=d.T.object)]), + withParameters(parameters): { parameters: parameters }, + '#withParametersMixin':: d.fn(help='Parameters holds the parameters for the provisioner that should create volumes of this storage class.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='parameters', type=d.T.object)]), + withParametersMixin(parameters): { parameters+: parameters }, + '#withProvisioner':: d.fn(help='Provisioner indicates the type of the provisioner.', args=[d.arg(name='provisioner', type=d.T.string)]), + withProvisioner(provisioner): { provisioner: provisioner }, + '#withReclaimPolicy':: d.fn(help='Dynamically provisioned PersistentVolumes of this storage class are created with this reclaimPolicy. Defaults to Delete.', args=[d.arg(name='reclaimPolicy', type=d.T.string)]), + withReclaimPolicy(reclaimPolicy): { reclaimPolicy: reclaimPolicy }, + '#withVolumeBindingMode':: d.fn(help='VolumeBindingMode indicates how PersistentVolumeClaims should be provisioned and bound. When unset, VolumeBindingImmediate is used. This field is only honored by servers that enable the VolumeScheduling feature.', args=[d.arg(name='volumeBindingMode', type=d.T.string)]), + withVolumeBindingMode(volumeBindingMode): { volumeBindingMode: volumeBindingMode }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachment.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachment.libsonnet new file mode 100644 index 00000000000..6e35fb1fd4a --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachment.libsonnet @@ -0,0 +1,479 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachment', url='', help='VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node.\n\nVolumeAttachment objects are non-namespaced.'), + '#metadata':: d.obj(help='ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create.'), + metadata: { + '#withAnnotations':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request.', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only.', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='A sequence number representing a specific generation of the desired state. Populated by the system. Read-only.', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withManagedFields':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFields(managedFields): { metadata+: { managedFields: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withManagedFieldsMixin':: d.fn(help="ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like 'ci-cd'. The set of fields is always in the version that the workflow used when modifying the object.\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='managedFields', type=d.T.array)]), + withManagedFieldsMixin(managedFields): { metadata+: { managedFields+: if std.isArray(v=managedFields) then managedFields else [managedFields] } }, + '#withName':: d.fn(help='Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release.', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Volumeattachment', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'storage.k8s.io/v1beta1', + kind: 'VolumeAttachment', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + spec: { + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { spec+: { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { spec+: { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { spec+: { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { spec+: { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { spec+: { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { spec+: { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { spec+: { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { spec+: { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { spec+: { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { spec+: { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { spec+: { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { spec+: { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity: capacity } } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { spec+: { source+: { inlineVolumeSpec+: { capacity+: capacity } } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { spec+: { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { spec+: { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { spec+: { source+: { persistentVolumeName: persistentVolumeName } } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { spec+: { attacher: attacher } }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { spec+: { nodeName: nodeName } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSource.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSource.libsonnet new file mode 100644 index 00000000000..65b29e14e10 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSource.libsonnet @@ -0,0 +1,419 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSource', url='', help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { azureDisk+: { kind: kind } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { cephfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { cephfs+: { user: user } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { cinder+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { claimRef+: { kind: kind } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { claimRef+: { name: name } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { claimRef+: { uid: uid } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { csi+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { csi+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { fc+: { fsType: fsType } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { fc+: { lun: lun } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { inlineVolumeSpec+: { flexVolume+: { driver: driver } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { inlineVolumeSpec+: { flexVolume+: { options: options } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { inlineVolumeSpec+: { flexVolume+: { options+: options } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { glusterfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { hostPath+: { path: path } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { inlineVolumeSpec+: { hostPath+: { type: type } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { inlineVolumeSpec+: { iscsi+: { lun: lun } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { 'local'+: { path: path } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { inlineVolumeSpec+: { nfs+: { path: path } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { inlineVolumeSpec+: { nfs+: { server: server } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { inlineVolumeSpec+: { quobyte+: { group: group } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { inlineVolumeSpec+: { quobyte+: { registry: registry } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { quobyte+: { user: user } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { inlineVolumeSpec+: { quobyte+: { volume: volume } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { rbd+: { fsType: fsType } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { inlineVolumeSpec+: { rbd+: { image: image } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { inlineVolumeSpec+: { rbd+: { keyring: keyring } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { inlineVolumeSpec+: { rbd+: { pool: pool } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { inlineVolumeSpec+: { rbd+: { user: user } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { inlineVolumeSpec+: { scaleIO+: { system: system } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { storageos+: { fsType: fsType } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { inlineVolumeSpec+: { capacity: capacity } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { inlineVolumeSpec+: { capacity+: capacity } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { inlineVolumeSpec+: { storageClassName: storageClassName } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { inlineVolumeSpec+: { volumeMode: volumeMode } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { persistentVolumeName: persistentVolumeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSpec.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSpec.libsonnet new file mode 100644 index 00000000000..29f116537cf --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentSpec.libsonnet @@ -0,0 +1,426 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentSpec', url='', help='VolumeAttachmentSpec is the specification of a VolumeAttachment request.'), + '#source':: d.obj(help='VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set.'), + source: { + '#inlineVolumeSpec':: d.obj(help='PersistentVolumeSpec is the specification of a persistent volume.'), + inlineVolumeSpec: { + '#awsElasticBlockStore':: d.obj(help='Represents a Persistent Disk resource in AWS.\n\nAn AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling.'), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty).', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { partition: partition } } } }, + '#withReadOnly':: d.fn(help='Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { awsElasticBlockStore+: { volumeID: volumeID } } } }, + }, + '#azureDisk':: d.obj(help='AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod.'), + azureDisk: { + '#withCachingMode':: d.fn(help='Host Caching mode: None, Read Only, Read Write.', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { source+: { inlineVolumeSpec+: { azureDisk+: { cachingMode: cachingMode } } } }, + '#withDiskName':: d.fn(help='The Name of the data disk in the blob storage', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { source+: { inlineVolumeSpec+: { azureDisk+: { diskName: diskName } } } }, + '#withDiskURI':: d.fn(help='The URI the data disk in the blob storage', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { source+: { inlineVolumeSpec+: { azureDisk+: { diskURI: diskURI } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { azureDisk+: { fsType: fsType } } } }, + '#withKind':: d.fn(help='Expected values Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { azureDisk+: { kind: kind } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureDisk+: { readOnly: readOnly } } } }, + }, + '#azureFile':: d.obj(help='AzureFile represents an Azure File Service mount on the host and bind mount to the pod.'), + azureFile: { + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { azureFile+: { readOnly: readOnly } } } }, + '#withSecretName':: d.fn(help='the name of secret that contains Azure Storage Account Name and Key', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { source+: { inlineVolumeSpec+: { azureFile+: { secretName: secretName } } } }, + '#withSecretNamespace':: d.fn(help='the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod', args=[d.arg(name='secretNamespace', type=d.T.string)]), + withSecretNamespace(secretNamespace): { source+: { inlineVolumeSpec+: { azureFile+: { secretNamespace: secretNamespace } } } }, + '#withShareName':: d.fn(help='Share Name', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { source+: { inlineVolumeSpec+: { azureFile+: { shareName: shareName } } } }, + }, + '#cephfs':: d.obj(help='Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling.'), + cephfs: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cephfs+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withMonitors':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPath':: d.fn(help='Optional: Used as the mounted root, rather than the full Ceph tree, default is /', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { cephfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cephfs+: { readOnly: readOnly } } } }, + '#withSecretFile':: d.fn(help='Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { source+: { inlineVolumeSpec+: { cephfs+: { secretFile: secretFile } } } }, + '#withUser':: d.fn(help='Optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { cephfs+: { user: user } } } }, + }, + '#cinder':: d.obj(help='Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling.'), + cinder: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { cinder+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { cinder+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { cinder+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='volume id used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { cinder+: { volumeID: volumeID } } } }, + }, + '#claimRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + claimRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { claimRef+: { apiVersion: apiVersion } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { claimRef+: { fieldPath: fieldPath } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { claimRef+: { kind: kind } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { claimRef+: { name: name } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { claimRef+: { namespace: namespace } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { claimRef+: { resourceVersion: resourceVersion } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { claimRef+: { uid: uid } } } }, + }, + '#csi':: d.obj(help='Represents storage that is managed by an external CSI volume driver (Beta feature)'), + csi: { + '#controllerExpandSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerExpandSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerExpandSecretRef+: { namespace: namespace } } } } }, + }, + '#controllerPublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + controllerPublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { controllerPublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodePublishSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodePublishSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodePublishSecretRef+: { namespace: namespace } } } } }, + }, + '#nodeStageSecretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + nodeStageSecretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { csi+: { nodeStageSecretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume. Required.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { csi+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs".', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { csi+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write).', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { csi+: { readOnly: readOnly } } } }, + '#withVolumeAttributes':: d.fn(help='Attributes of the volume to publish.', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes: volumeAttributes } } } }, + '#withVolumeAttributesMixin':: d.fn(help='Attributes of the volume to publish.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { source+: { inlineVolumeSpec+: { csi+: { volumeAttributes+: volumeAttributes } } } }, + '#withVolumeHandle':: d.fn(help='VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required.', args=[d.arg(name='volumeHandle', type=d.T.string)]), + withVolumeHandle(volumeHandle): { source+: { inlineVolumeSpec+: { csi+: { volumeHandle: volumeHandle } } } }, + }, + '#fc':: d.obj(help='Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling.'), + fc: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { fc+: { fsType: fsType } } } }, + '#withLun':: d.fn(help='Optional: FC target lun number', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { fc+: { lun: lun } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { fc+: { readOnly: readOnly } } } }, + '#withTargetWWNs':: d.fn(help='Optional: FC target worldwide names (WWNs)', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withTargetWWNsMixin':: d.fn(help='Optional: FC target worldwide names (WWNs)\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { source+: { inlineVolumeSpec+: { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } } } }, + '#withWwids':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + '#withWwidsMixin':: d.fn(help='Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { source+: { inlineVolumeSpec+: { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } } } }, + }, + '#flexVolume':: d.obj(help='FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin.'), + flexVolume: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { flexVolume+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withDriver':: d.fn(help='Driver is the name of the driver to use for this volume.', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { source+: { inlineVolumeSpec+: { flexVolume+: { driver: driver } } } }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { flexVolume+: { fsType: fsType } } } }, + '#withOptions':: d.fn(help='Optional: Extra command options if any.', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options: options } } } }, + '#withOptionsMixin':: d.fn(help='Optional: Extra command options if any.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { source+: { inlineVolumeSpec+: { flexVolume+: { options+: options } } } }, + '#withReadOnly':: d.fn(help='Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { flexVolume+: { readOnly: readOnly } } } }, + }, + '#flocker':: d.obj(help='Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling.'), + flocker: { + '#withDatasetName':: d.fn(help='Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { source+: { inlineVolumeSpec+: { flocker+: { datasetName: datasetName } } } }, + '#withDatasetUUID':: d.fn(help='UUID of the dataset. This is unique identifier of a Flocker dataset', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { source+: { inlineVolumeSpec+: { flocker+: { datasetUUID: datasetUUID } } } }, + }, + '#gcePersistentDisk':: d.obj(help='Represents a Persistent Disk resource in Google Compute Engine.\n\nA GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling.'), + gcePersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { fsType: fsType } } } }, + '#withPartition':: d.fn(help='The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { partition: partition } } } }, + '#withPdName':: d.fn(help='Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { pdName: pdName } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { gcePersistentDisk+: { readOnly: readOnly } } } }, + }, + '#glusterfs':: d.obj(help='Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling.'), + glusterfs: { + '#withEndpoints':: d.fn(help='EndpointsName is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { source+: { inlineVolumeSpec+: { glusterfs+: { endpoints: endpoints } } } }, + '#withEndpointsNamespace':: d.fn(help='EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='endpointsNamespace', type=d.T.string)]), + withEndpointsNamespace(endpointsNamespace): { source+: { inlineVolumeSpec+: { glusterfs+: { endpointsNamespace: endpointsNamespace } } } }, + '#withPath':: d.fn(help='Path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { glusterfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { glusterfs+: { readOnly: readOnly } } } }, + }, + '#hostPath':: d.obj(help='Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling.'), + hostPath: { + '#withPath':: d.fn(help='Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { hostPath+: { path: path } } } }, + '#withType':: d.fn(help='Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { source+: { inlineVolumeSpec+: { hostPath+: { type: type } } } }, + }, + '#iscsi':: d.obj(help='ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling.'), + iscsi: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { iscsi+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='whether support iSCSI Discovery CHAP authentication', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } } } }, + '#withChapAuthSession':: d.fn(help='whether support iSCSI Session CHAP authentication', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { source+: { inlineVolumeSpec+: { iscsi+: { chapAuthSession: chapAuthSession } } } }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { iscsi+: { fsType: fsType } } } }, + '#withInitiatorName':: d.fn(help='Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection.', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { source+: { inlineVolumeSpec+: { iscsi+: { initiatorName: initiatorName } } } }, + '#withIqn':: d.fn(help='Target iSCSI Qualified Name.', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { source+: { inlineVolumeSpec+: { iscsi+: { iqn: iqn } } } }, + '#withIscsiInterface':: d.fn(help="iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { source+: { inlineVolumeSpec+: { iscsi+: { iscsiInterface: iscsiInterface } } } }, + '#withLun':: d.fn(help='iSCSI Target Lun number.', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { source+: { inlineVolumeSpec+: { iscsi+: { lun: lun } } } }, + '#withPortals':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withPortalsMixin':: d.fn(help='iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { source+: { inlineVolumeSpec+: { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { iscsi+: { readOnly: readOnly } } } }, + '#withTargetPortal':: d.fn(help='iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260).', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { source+: { inlineVolumeSpec+: { iscsi+: { targetPortal: targetPortal } } } }, + }, + '#local':: d.obj(help='Local represents directly-attached storage with node affinity (Beta feature)'), + 'local': { + '#withFsType':: d.fn(help='Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { 'local'+: { fsType: fsType } } } }, + '#withPath':: d.fn(help='The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...).', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { 'local'+: { path: path } } } }, + }, + '#nfs':: d.obj(help='Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling.'), + nfs: { + '#withPath':: d.fn(help='Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { source+: { inlineVolumeSpec+: { nfs+: { path: path } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { nfs+: { readOnly: readOnly } } } }, + '#withServer':: d.fn(help='Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { source+: { inlineVolumeSpec+: { nfs+: { server: server } } } }, + }, + '#nodeAffinity':: d.obj(help='VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from.'), + nodeAffinity: { + '#required':: d.obj(help='A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms.'), + required: { + '#withNodeSelectorTerms':: d.fn(help='Required. A list of node selector terms. The terms are ORed.', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='Required. A list of node selector terms. The terms are ORed.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { source+: { inlineVolumeSpec+: { nodeAffinity+: { required+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + }, + '#photonPersistentDisk':: d.obj(help='Represents a Photon Controller persistent disk resource.'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { fsType: fsType } } } }, + '#withPdID':: d.fn(help='ID that identifies Photon Controller persistent disk', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { source+: { inlineVolumeSpec+: { photonPersistentDisk+: { pdID: pdID } } } }, + }, + '#portworxVolume':: d.obj(help='PortworxVolumeSource represents a Portworx volume resource.'), + portworxVolume: { + '#withFsType':: d.fn(help='FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { portworxVolume+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { portworxVolume+: { readOnly: readOnly } } } }, + '#withVolumeID':: d.fn(help='VolumeID uniquely identifies a Portworx volume', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { source+: { inlineVolumeSpec+: { portworxVolume+: { volumeID: volumeID } } } }, + }, + '#quobyte':: d.obj(help='Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling.'), + quobyte: { + '#withGroup':: d.fn(help='Group to map volume access to Default is no group', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { source+: { inlineVolumeSpec+: { quobyte+: { group: group } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { quobyte+: { readOnly: readOnly } } } }, + '#withRegistry':: d.fn(help='Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { source+: { inlineVolumeSpec+: { quobyte+: { registry: registry } } } }, + '#withTenant':: d.fn(help='Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { source+: { inlineVolumeSpec+: { quobyte+: { tenant: tenant } } } }, + '#withUser':: d.fn(help='User to map volume access to Defaults to serivceaccount user', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { quobyte+: { user: user } } } }, + '#withVolume':: d.fn(help='Volume is a string that references an already created Quobyte volume by name.', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { source+: { inlineVolumeSpec+: { quobyte+: { volume: volume } } } }, + }, + '#rbd':: d.obj(help='Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling.'), + rbd: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { rbd+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { rbd+: { fsType: fsType } } } }, + '#withImage':: d.fn(help='The rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { source+: { inlineVolumeSpec+: { rbd+: { image: image } } } }, + '#withKeyring':: d.fn(help='Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { source+: { inlineVolumeSpec+: { rbd+: { keyring: keyring } } } }, + '#withMonitors':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withMonitorsMixin':: d.fn(help='A collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { source+: { inlineVolumeSpec+: { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } } } }, + '#withPool':: d.fn(help='The rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { source+: { inlineVolumeSpec+: { rbd+: { pool: pool } } } }, + '#withReadOnly':: d.fn(help='ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { rbd+: { readOnly: readOnly } } } }, + '#withUser':: d.fn(help='The rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { source+: { inlineVolumeSpec+: { rbd+: { user: user } } } }, + }, + '#scaleIO':: d.obj(help='ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume'), + scaleIO: { + '#secretRef':: d.obj(help='SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace'), + secretRef: { + '#withName':: d.fn(help='Name is unique within a namespace to reference a secret resource.', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace defines the space within which the secret name must be unique.', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { scaleIO+: { secretRef+: { namespace: namespace } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { scaleIO+: { fsType: fsType } } } }, + '#withGateway':: d.fn(help='The host address of the ScaleIO API Gateway.', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { source+: { inlineVolumeSpec+: { scaleIO+: { gateway: gateway } } } }, + '#withProtectionDomain':: d.fn(help='The name of the ScaleIO Protection Domain for the configured storage.', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { source+: { inlineVolumeSpec+: { scaleIO+: { protectionDomain: protectionDomain } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { scaleIO+: { readOnly: readOnly } } } }, + '#withSslEnabled':: d.fn(help='Flag to enable/disable SSL communication with Gateway, default false', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { source+: { inlineVolumeSpec+: { scaleIO+: { sslEnabled: sslEnabled } } } }, + '#withStorageMode':: d.fn(help='Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned.', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { source+: { inlineVolumeSpec+: { scaleIO+: { storageMode: storageMode } } } }, + '#withStoragePool':: d.fn(help='The ScaleIO Storage Pool associated with the protection domain.', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { source+: { inlineVolumeSpec+: { scaleIO+: { storagePool: storagePool } } } }, + '#withSystem':: d.fn(help='The name of the storage system as configured in ScaleIO.', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { source+: { inlineVolumeSpec+: { scaleIO+: { system: system } } } }, + '#withVolumeName':: d.fn(help='The name of a volume already created in the ScaleIO system that is associated with this volume source.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { scaleIO+: { volumeName: volumeName } } } }, + }, + '#storageos':: d.obj(help='Represents a StorageOS persistent volume resource.'), + storageos: { + '#secretRef':: d.obj(help='ObjectReference contains enough information to let you inspect or modify the referred object.'), + secretRef: { + '#withApiVersion':: d.fn(help='API version of the referent.', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { apiVersion: apiVersion } } } } }, + '#withFieldPath':: d.fn(help='If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object.', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { fieldPath: fieldPath } } } } }, + '#withKind':: d.fn(help='Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help='Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { namespace: namespace } } } } }, + '#withResourceVersion':: d.fn(help='Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { resourceVersion: resourceVersion } } } } }, + '#withUid':: d.fn(help='UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { source+: { inlineVolumeSpec+: { storageos+: { secretRef+: { uid: uid } } } } }, + }, + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { storageos+: { fsType: fsType } } } }, + '#withReadOnly':: d.fn(help='Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts.', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { source+: { inlineVolumeSpec+: { storageos+: { readOnly: readOnly } } } }, + '#withVolumeName':: d.fn(help='VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace.', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { source+: { inlineVolumeSpec+: { storageos+: { volumeName: volumeName } } } }, + '#withVolumeNamespace':: d.fn(help="VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to 'default' if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { source+: { inlineVolumeSpec+: { storageos+: { volumeNamespace: volumeNamespace } } } }, + }, + '#vsphereVolume':: d.obj(help='Represents a vSphere volume resource.'), + vsphereVolume: { + '#withFsType':: d.fn(help='Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified.', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { source+: { inlineVolumeSpec+: { vsphereVolume+: { fsType: fsType } } } }, + '#withStoragePolicyID':: d.fn(help='Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName.', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyID: storagePolicyID } } } }, + '#withStoragePolicyName':: d.fn(help='Storage Policy Based Management (SPBM) profile name.', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { source+: { inlineVolumeSpec+: { vsphereVolume+: { storagePolicyName: storagePolicyName } } } }, + '#withVolumePath':: d.fn(help='Path that identifies vSphere volume vmdk', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { source+: { inlineVolumeSpec+: { vsphereVolume+: { volumePath: volumePath } } } }, + }, + '#withAccessModes':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { source+: { inlineVolumeSpec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withAccessModesMixin':: d.fn(help='AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { source+: { inlineVolumeSpec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } }, + '#withCapacity':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacity(capacity): { source+: { inlineVolumeSpec+: { capacity: capacity } } }, + '#withCapacityMixin':: d.fn(help="A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='capacity', type=d.T.object)]), + withCapacityMixin(capacity): { source+: { inlineVolumeSpec+: { capacity+: capacity } } }, + '#withMountOptions':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptions(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withMountOptionsMixin':: d.fn(help='A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='mountOptions', type=d.T.array)]), + withMountOptionsMixin(mountOptions): { source+: { inlineVolumeSpec+: { mountOptions+: if std.isArray(v=mountOptions) then mountOptions else [mountOptions] } } }, + '#withPersistentVolumeReclaimPolicy':: d.fn(help='What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming', args=[d.arg(name='persistentVolumeReclaimPolicy', type=d.T.string)]), + withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy): { source+: { inlineVolumeSpec+: { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy } } }, + '#withStorageClassName':: d.fn(help='Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass.', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { source+: { inlineVolumeSpec+: { storageClassName: storageClassName } } }, + '#withVolumeMode':: d.fn(help='volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec.', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { source+: { inlineVolumeSpec+: { volumeMode: volumeMode } } }, + }, + '#withPersistentVolumeName':: d.fn(help='Name of the persistent volume to attach.', args=[d.arg(name='persistentVolumeName', type=d.T.string)]), + withPersistentVolumeName(persistentVolumeName): { source+: { persistentVolumeName: persistentVolumeName } }, + }, + '#withAttacher':: d.fn(help='Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName().', args=[d.arg(name='attacher', type=d.T.string)]), + withAttacher(attacher): { attacher: attacher }, + '#withNodeName':: d.fn(help='The node that the volume should be attached to.', args=[d.arg(name='nodeName', type=d.T.string)]), + withNodeName(nodeName): { nodeName: nodeName }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentStatus.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentStatus.libsonnet new file mode 100644 index 00000000000..7d14c51ec30 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeAttachmentStatus.libsonnet @@ -0,0 +1,26 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeAttachmentStatus', url='', help='VolumeAttachmentStatus is the status of a VolumeAttachment request.'), + '#attachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + attachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { attachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { attachError+: { time: time } }, + }, + '#detachError':: d.obj(help='VolumeError captures an error encountered during a volume operation.'), + detachError: { + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { detachError+: { message: message } }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { detachError+: { time: time } }, + }, + '#withAttached':: d.fn(help='Indicates the volume is successfully attached. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attached', type=d.T.boolean)]), + withAttached(attached): { attached: attached }, + '#withAttachmentMetadata':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadata(attachmentMetadata): { attachmentMetadata: attachmentMetadata }, + '#withAttachmentMetadataMixin':: d.fn(help='Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher.\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='attachmentMetadata', type=d.T.object)]), + withAttachmentMetadataMixin(attachmentMetadata): { attachmentMetadata+: attachmentMetadata }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeError.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeError.libsonnet new file mode 100644 index 00000000000..aea460dc2e1 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeError.libsonnet @@ -0,0 +1,10 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeError', url='', help='VolumeError captures an error encountered during a volume operation.'), + '#withMessage':: d.fn(help='String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information.', args=[d.arg(name='message', type=d.T.string)]), + withMessage(message): { message: message }, + '#withTime':: d.fn(help='Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers.', args=[d.arg(name='time', type=d.T.string)]), + withTime(time): { time: time }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeNodeResources.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeNodeResources.libsonnet new file mode 100644 index 00000000000..6fe6f33a0f5 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/_gen/storage/v1beta1/volumeNodeResources.libsonnet @@ -0,0 +1,8 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='volumeNodeResources', url='', help='VolumeNodeResources is a set of resource limits for scheduling of volumes.'), + '#withCount':: d.fn(help='Maximum number of unique volumes managed by the CSI driver that can be used on a node. A volume that is both attached and mounted on a node is considered to be used once, not twice. The same rule applies for a unique volume that is shared among multiple pods on the same node. If this field is nil, then the supported number of volumes on this node is unbounded.', args=[d.arg(name='count', type=d.T.integer)]), + withCount(count): { count: count }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/gen.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/gen.libsonnet new file mode 100644 index 00000000000..3121b502ec3 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/gen.libsonnet @@ -0,0 +1,25 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='k', url='github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet', help='Generated Jsonnet library for Kubernetes v1.19'), + admissionregistration:: (import '_gen/admissionregistration/main.libsonnet'), + apiregistration:: (import '_gen/apiregistration/main.libsonnet'), + apps:: (import '_gen/apps/main.libsonnet'), + authentication:: (import '_gen/authentication/main.libsonnet'), + authorization:: (import '_gen/authorization/main.libsonnet'), + autoscaling:: (import '_gen/autoscaling/main.libsonnet'), + batch:: (import '_gen/batch/main.libsonnet'), + certificates:: (import '_gen/certificates/main.libsonnet'), + coordination:: (import '_gen/coordination/main.libsonnet'), + core:: (import '_gen/core/main.libsonnet'), + discovery:: (import '_gen/discovery/main.libsonnet'), + events:: (import '_gen/events/main.libsonnet'), + extensions:: (import '_gen/extensions/main.libsonnet'), + flowcontrol:: (import '_gen/flowcontrol/main.libsonnet'), + networking:: (import '_gen/networking/main.libsonnet'), + node:: (import '_gen/node/main.libsonnet'), + policy:: (import '_gen/policy/main.libsonnet'), + rbac:: (import '_gen/rbac/main.libsonnet'), + scheduling:: (import '_gen/scheduling/main.libsonnet'), + settings:: (import '_gen/settings/main.libsonnet'), + storage:: (import '_gen/storage/main.libsonnet'), +} diff --git a/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet new file mode 100644 index 00000000000..2404d5241a9 --- /dev/null +++ b/example/tk/vendor/github.com/jsonnet-libs/k8s-libsonnet/1.19/main.libsonnet @@ -0,0 +1 @@ +(import 'gen.libsonnet') + (import '_custom/apps.libsonnet') + (import '_custom/batch.libsonnet') + (import '_custom/core.libsonnet') + (import '_custom/list.libsonnet') + (import '_custom/mapContainers.libsonnet') + (import '_custom/rbac.libsonnet') + (import '_custom/volumeMounts.libsonnet') diff --git a/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k.libsonnet b/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k.libsonnet deleted file mode 100644 index c50661a4e87..00000000000 --- a/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k.libsonnet +++ /dev/null @@ -1,123 +0,0 @@ -local k8s = import 'k8s.libsonnet'; -local fn = { - mapContainers(f):: { - local podContainers = super.spec.template.spec.containers, - spec+: { - template+: { - spec+: { - containers: std.map(f, podContainers), - }, - }, - }, - }, - mapContainersWithName(names, f):: - local nameSet = if std.type(names) == 'array' then std.set(names) else std.set([names]); - local inNameSet(name) = std.length(std.setInter(nameSet, std.set([name]))) > 0; - - self.mapContainers(function(c) if std.objectHas(c, 'name') && inNameSet(c.name) then f(c) else c), -}; - -k8s + { - apps:: k8s.apps + { - v1:: k8s.apps.v1 + { - daemonSet:: k8s.apps.v1.daemonSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - deployment:: k8s.apps.v1.deployment + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - replicaSet:: k8s.apps.v1.replicaSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - statefulSet:: k8s.apps.v1.statefulSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - v1beta1:: k8s.apps.v1beta1 + { - deployment:: k8s.apps.v1beta1.deployment + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - statefulSet:: k8s.apps.v1beta1.statefulSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - v1beta2:: k8s.apps.v1beta2 + { - daemonSet:: k8s.apps.v1beta2.daemonSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - deployment:: k8s.apps.v1beta2.deployment + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - replicaSet:: k8s.apps.v1beta2.replicaSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - statefulSet:: k8s.apps.v1beta2.statefulSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - }, - batch:: k8s.batch + { - v1:: k8s.batch.v1 + { - job:: k8s.batch.v1.job + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - v1beta1:: k8s.batch.v1beta1 + { - cronJob:: k8s.batch.v1beta1.cronJob + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - }, - core:: k8s.core + { - v1:: k8s.core.v1 + { - list:: { - new(items):: { - apiVersion: 'v1', - } + { - kind: 'List', - } + self.items(items), - items(items):: if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - }, - pod:: k8s.core.v1.pod + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - podTemplate:: k8s.core.v1.podTemplate + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - replicationController:: k8s.core.v1.replicationController + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - }, - extensions:: k8s.extensions + { - v1beta1:: k8s.extensions.v1beta1 + { - daemonSet:: k8s.extensions.v1beta1.daemonSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - deployment:: k8s.extensions.v1beta1.deployment + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - replicaSet:: k8s.extensions.v1beta1.replicaSet + { - mapContainers(f):: fn.mapContainers(f), - mapContainersWithName(names, f):: fn.mapContainersWithName(names, f), - }, - }, - }, -} \ No newline at end of file diff --git a/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k8s.libsonnet b/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k8s.libsonnet deleted file mode 100644 index 2400f87f67b..00000000000 --- a/example/tk/vendor/github.com/ksonnet/ksonnet-lib/ksonnet.beta.4/k8s.libsonnet +++ /dev/null @@ -1,30584 +0,0 @@ -{ - __ksonnet: { - checksum: 'ae84879f3189d1d6cfa3e1cf79356fbdc05649900e71369875ba4a45a408863f', - kubernetesVersion: '1.14.0', - }, - admissionregistration:: { - v1beta1:: { - local apiVersion = { apiVersion: 'admissionregistration.k8s.io/v1beta1' }, - // MutatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and may change the object. - mutatingWebhookConfiguration:: { - local kind = { kind: 'MutatingWebhookConfiguration' }, - new():: apiVersion + kind, - // Webhooks is a list of webhooks and the affected resources and operations. - withWebhooks(webhooks):: self + if std.type(webhooks) == 'array' then { webhooks: webhooks } else { webhooks: [webhooks] }, - // Webhooks is a list of webhooks and the affected resources and operations. - withWebhooksMixin(webhooks):: self + if std.type(webhooks) == 'array' then { webhooks+: webhooks } else { webhooks+: [webhooks] }, - webhooksType:: hidden.admissionregistration.v1beta1.webhook, - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // ValidatingWebhookConfiguration describes the configuration of and admission webhook that accept or reject and object without changing it. - validatingWebhookConfiguration:: { - local kind = { kind: 'ValidatingWebhookConfiguration' }, - new():: apiVersion + kind, - // Webhooks is a list of webhooks and the affected resources and operations. - withWebhooks(webhooks):: self + if std.type(webhooks) == 'array' then { webhooks: webhooks } else { webhooks: [webhooks] }, - // Webhooks is a list of webhooks and the affected resources and operations. - withWebhooksMixin(webhooks):: self + if std.type(webhooks) == 'array' then { webhooks+: webhooks } else { webhooks+: [webhooks] }, - webhooksType:: hidden.admissionregistration.v1beta1.webhook, - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - }, - }, - apiextensions:: { - v1beta1:: { - local apiVersion = { apiVersion: 'apiextensions.k8s.io/v1beta1' }, - // CustomResourceDefinition represents a resource that should be exposed on the API server. Its name MUST be in the format <.spec.name>.<.spec.group>. - customResourceDefinition:: { - local kind = { kind: 'CustomResourceDefinition' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec describes how the user wants the resources to appear - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Optional, the global columns for all versions. Top-level and per-version columns are mutually exclusive. - withAdditionalPrinterColumns(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then __specMixin({ additionalPrinterColumns: additionalPrinterColumns }) else __specMixin({ additionalPrinterColumns: [additionalPrinterColumns] }), - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Optional, the global columns for all versions. Top-level and per-version columns are mutually exclusive. - withAdditionalPrinterColumnsMixin(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then __specMixin({ additionalPrinterColumns+: additionalPrinterColumns }) else __specMixin({ additionalPrinterColumns+: [additionalPrinterColumns] }), - additionalPrinterColumnsType:: hidden.apiextensions.v1beta1.customResourceColumnDefinition, - // `conversion` defines conversion settings for the CRD. - conversion:: { - local __conversionMixin(conversion) = __specMixin({ conversion+: conversion }), - mixinInstance(conversion):: __conversionMixin(conversion), - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersions(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then __conversionMixin({ conversionReviewVersions: conversionReviewVersions }) else __conversionMixin({ conversionReviewVersions: [conversionReviewVersions] }), - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersionsMixin(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then __conversionMixin({ conversionReviewVersions+: conversionReviewVersions }) else __conversionMixin({ conversionReviewVersions+: [conversionReviewVersions] }), - // `strategy` specifies the conversion strategy. Allowed values are: - `None`: The converter only change the apiVersion and would not touch any other field in the CR. - `Webhook`: API Server will call to an external webhook to do the conversion. Additional information is needed for this option. - withStrategy(strategy):: self + __conversionMixin({ strategy: strategy }), - // `webhookClientConfig` is the instructions for how to call the webhook if strategy is `Webhook`. This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. - webhookClientConfig:: { - local __webhookClientConfigMixin(webhookClientConfig) = __conversionMixin({ webhookClientConfig+: webhookClientConfig }), - mixinInstance(webhookClientConfig):: __webhookClientConfigMixin(webhookClientConfig), - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __webhookClientConfigMixin({ caBundle: caBundle }), - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = __webhookClientConfigMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.apiextensions.v1beta1.serviceReference, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + __webhookClientConfigMixin({ url: url }), - }, - webhookClientConfigType:: hidden.apiextensions.v1beta1.webhookClientConfig, - }, - conversionType:: hidden.apiextensions.v1beta1.customResourceConversion, - // Group is the group this resource belongs in - withGroup(group):: self + __specMixin({ group: group }), - // Names are the names used to describe this custom resource - names:: { - local __namesMixin(names) = __specMixin({ names+: names }), - mixinInstance(names):: __namesMixin(names), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategories(categories):: self + if std.type(categories) == 'array' then __namesMixin({ categories: categories }) else __namesMixin({ categories: [categories] }), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategoriesMixin(categories):: self + if std.type(categories) == 'array' then __namesMixin({ categories+: categories }) else __namesMixin({ categories+: [categories] }), - // Kind is the serialized kind of the resource. It is normally CamelCase and singular. - withKind(kind):: self + __namesMixin({ kind: kind }), - // ListKind is the serialized kind of the list for this resource. Defaults to List. - withListKind(listKind):: self + __namesMixin({ listKind: listKind }), - // Plural is the plural name of the resource to serve. It must match the name of the CustomResourceDefinition-registration too: plural.group and it must be all lowercase. - withPlural(plural):: self + __namesMixin({ plural: plural }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNames(shortNames):: self + if std.type(shortNames) == 'array' then __namesMixin({ shortNames: shortNames }) else __namesMixin({ shortNames: [shortNames] }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNamesMixin(shortNames):: self + if std.type(shortNames) == 'array' then __namesMixin({ shortNames+: shortNames }) else __namesMixin({ shortNames+: [shortNames] }), - // Singular is the singular name of the resource. It must be all lowercase Defaults to lowercased - withSingular(singular):: self + __namesMixin({ singular: singular }), - }, - namesType:: hidden.apiextensions.v1beta1.customResourceDefinitionNames, - // Scope indicates whether this resource is cluster or namespace scoped. Default is namespaced - withScope(scope):: self + __specMixin({ scope: scope }), - // Subresources describes the subresources for CustomResource Optional, the global subresources for all versions. Top-level and per-version subresources are mutually exclusive. - subresources:: { - local __subresourcesMixin(subresources) = __specMixin({ subresources+: subresources }), - mixinInstance(subresources):: __subresourcesMixin(subresources), - // Scale denotes the scale subresource for CustomResources - scale:: { - local __scaleMixin(scale) = __subresourcesMixin({ scale+: scale }), - mixinInstance(scale):: __scaleMixin(scale), - // LabelSelectorPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Selector. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. Must be set to work with HPA. If there is no value under the given path in the CustomResource, the status label selector value in the /scale subresource will default to the empty string. - withLabelSelectorPath(labelSelectorPath):: self + __scaleMixin({ labelSelectorPath: labelSelectorPath }), - // SpecReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Spec.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .spec. If there is no value under the given path in the CustomResource, the /scale subresource will return an error on GET. - withSpecReplicasPath(specReplicasPath):: self + __scaleMixin({ specReplicasPath: specReplicasPath }), - // StatusReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. If there is no value under the given path in the CustomResource, the status replica value in the /scale subresource will default to 0. - withStatusReplicasPath(statusReplicasPath):: self + __scaleMixin({ statusReplicasPath: statusReplicasPath }), - }, - scaleType:: hidden.apiextensions.v1beta1.customResourceSubresourceScale, - }, - subresourcesType:: hidden.apiextensions.v1beta1.customResourceSubresources, - // Validation describes the validation methods for CustomResources Optional, the global validation schema for all versions. Top-level and per-version schemas are mutually exclusive. - validation:: { - local __validationMixin(validation) = __specMixin({ validation+: validation }), - mixinInstance(validation):: __validationMixin(validation), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3Schema(openApiV3Schema):: self + __validationMixin({ openAPIV3Schema: openApiV3Schema }), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3SchemaMixin(openApiV3Schema):: self + __validationMixin({ openAPIV3Schema+: openApiV3Schema }), - openAPIV3SchemaType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - }, - validationType:: hidden.apiextensions.v1beta1.customResourceValidation, - // Version is the version this resource belongs in Should be always first item in Versions field if provided. Optional, but at least one of Version or Versions must be set. Deprecated: Please use `Versions`. - withVersion(version):: self + __specMixin({ version: version }), - // Versions is the list of all supported versions for this resource. If Version field is provided, this field is optional. Validation: All versions must use the same validation schema for now. i.e., top level Validation field is applied to all of these versions. Order: The version name will be used to compute the order. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersions(versions):: self + if std.type(versions) == 'array' then __specMixin({ versions: versions }) else __specMixin({ versions: [versions] }), - // Versions is the list of all supported versions for this resource. If Version field is provided, this field is optional. Validation: All versions must use the same validation schema for now. i.e., top level Validation field is applied to all of these versions. Order: The version name will be used to compute the order. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionsMixin(versions):: self + if std.type(versions) == 'array' then __specMixin({ versions+: versions }) else __specMixin({ versions+: [versions] }), - versionsType:: hidden.apiextensions.v1beta1.customResourceDefinitionVersion, - }, - specType:: hidden.apiextensions.v1beta1.customResourceDefinitionSpec, - }, - }, - }, - }, - apiregistration:: { - v1:: { - local apiVersion = { apiVersion: 'apiregistration.k8s.io/v1' }, - // APIService represents a server for a particular GroupVersion. Name must be "version.group". - apiService:: { - local kind = { kind: 'APIService' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec contains information for locating and communicating with a server - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __specMixin({ caBundle: caBundle }), - // Group is the API group name this server hosts - withGroup(group):: self + __specMixin({ group: group }), - // GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s - withGroupPriorityMinimum(groupPriorityMinimum):: self + __specMixin({ groupPriorityMinimum: groupPriorityMinimum }), - // InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead. - withInsecureSkipTlsVerify(insecureSkipTlsVerify):: self + __specMixin({ insecureSkipTLSVerify: insecureSkipTlsVerify }), - // Service is a reference to the service for this API server. It must communicate on port 443 If the Service is nil, that means the handling for the API groupversion is handled locally on this server. The call will simply delegate to the normal handler chain to be fulfilled. - service:: { - local __serviceMixin(service) = __specMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // Name is the name of the service - withName(name):: self + __serviceMixin({ name: name }), - // Namespace is the namespace of the service - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - }, - serviceType:: hidden.apiregistration.v1.serviceReference, - // Version is the API version this server hosts. For example, "v1" - withVersion(version):: self + __specMixin({ version: version }), - // VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionPriority(versionPriority):: self + __specMixin({ versionPriority: versionPriority }), - }, - specType:: hidden.apiregistration.v1.apiServiceSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'apiregistration.k8s.io/v1beta1' }, - // APIService represents a server for a particular GroupVersion. Name must be "version.group". - apiService:: { - local kind = { kind: 'APIService' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec contains information for locating and communicating with a server - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __specMixin({ caBundle: caBundle }), - // Group is the API group name this server hosts - withGroup(group):: self + __specMixin({ group: group }), - // GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s - withGroupPriorityMinimum(groupPriorityMinimum):: self + __specMixin({ groupPriorityMinimum: groupPriorityMinimum }), - // InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead. - withInsecureSkipTlsVerify(insecureSkipTlsVerify):: self + __specMixin({ insecureSkipTLSVerify: insecureSkipTlsVerify }), - // Service is a reference to the service for this API server. It must communicate on port 443 If the Service is nil, that means the handling for the API groupversion is handled locally on this server. The call will simply delegate to the normal handler chain to be fulfilled. - service:: { - local __serviceMixin(service) = __specMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // Name is the name of the service - withName(name):: self + __serviceMixin({ name: name }), - // Namespace is the namespace of the service - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - }, - serviceType:: hidden.apiregistration.v1beta1.serviceReference, - // Version is the API version this server hosts. For example, "v1" - withVersion(version):: self + __specMixin({ version: version }), - // VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionPriority(versionPriority):: self + __specMixin({ versionPriority: versionPriority }), - }, - specType:: hidden.apiregistration.v1beta1.apiServiceSpec, - }, - }, - }, - }, - apps:: { - v1:: { - local apiVersion = { apiVersion: 'apps/v1' }, - // ControllerRevision implements an immutable snapshot of state data. Clients are responsible for serializing and deserializing the objects that contain their internal state. Once a ControllerRevision has been successfully created, it can not be updated. The API Server will fail validation of all requests that attempt to mutate the Data field. ControllerRevisions may, however, be deleted. Note that, due to its use by both the DaemonSet and StatefulSet controllers for update and rollback, this object is beta. However, it may be subject to name and representation changes in future releases, and clients should not depend on its stability. It is primarily for internal use by controllers. - controllerRevision:: { - local kind = { kind: 'ControllerRevision' }, - new():: apiVersion + kind, - // Revision indicates the revision of the state represented by Data. - withRevision(revision):: self + { revision: revision }, - mixin:: { - // Data is the serialized representation of the state. - data:: { - local __dataMixin(data) = { data+: data }, - mixinInstance(data):: __dataMixin(data), - // Raw is the underlying serialization of this object. - withRaw(raw):: self + __dataMixin({ Raw: raw }), - }, - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // DaemonSet represents the configuration of a daemon set. - daemonSet:: { - local kind = { kind: 'DaemonSet' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The desired behavior of this daemon set. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1.daemonSetUpdateStrategy, - }, - specType:: hidden.apps.v1.daemonSetSpec, - }, - }, - // Deployment enables declarative updates for Pods and ReplicaSets. - deployment:: { - local kind = { kind: 'Deployment' }, - new(name='', replicas=1, containers='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - // Standard object metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the Deployment. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Indicates that the deployment is paused. - withPaused(paused):: self + __specMixin({ paused: paused }), - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + __specMixin({ progressDeadlineSeconds: progressDeadlineSeconds }), - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. It must match the pod template's labels. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = __specMixin({ strategy+: strategy }), - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.apps.v1.deploymentSpec, - }, - }, - // ReplicaSet ensures that a specified number of pod replicas are running at any given time. - replicaSet:: { - local kind = { kind: 'ReplicaSet' }, - new():: apiVersion + kind, - mixin:: { - // If the Labels of a ReplicaSet are empty, they are defaulted to be the same as the Pod(s) that the ReplicaSet manages. Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the specification of the desired behavior of the ReplicaSet. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // Selector is a label query over pods that should match the replica count. Label keys and values that must match in order to be controlled by this replica set. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.apps.v1.replicaSetSpec, - }, - }, - // StatefulSet represents a set of pods with consistent identities. Identities are defined as: - // - Network: A single stable DNS and hostname. - // - Storage: As many VolumeClaims as requested. - // The StatefulSet guarantees that a given network identity will always map to the same storage identity. - statefulSet:: { - local kind = { kind: 'StatefulSet' }, - new(name='', replicas=1, containers='', volumeClaims='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas).withVolumeClaimTemplates(volumeClaims) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the desired identities of pods in this set. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + __specMixin({ podManagementPolicy: podManagementPolicy }), - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // selector is a label query over pods that should match the replica count. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + __specMixin({ serviceName: serviceName }), - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1.statefulSetUpdateStrategy, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates: [volumeClaimTemplates] }), - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates+: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates+: [volumeClaimTemplates] }), - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - }, - specType:: hidden.apps.v1.statefulSetSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'apps/v1beta1' }, - // DEPRECATED - This group version of ControllerRevision is deprecated by apps/v1beta2/ControllerRevision. See the release notes for more information. ControllerRevision implements an immutable snapshot of state data. Clients are responsible for serializing and deserializing the objects that contain their internal state. Once a ControllerRevision has been successfully created, it can not be updated. The API Server will fail validation of all requests that attempt to mutate the Data field. ControllerRevisions may, however, be deleted. Note that, due to its use by both the DaemonSet and StatefulSet controllers for update and rollback, this object is beta. However, it may be subject to name and representation changes in future releases, and clients should not depend on its stability. It is primarily for internal use by controllers. - controllerRevision:: { - local kind = { kind: 'ControllerRevision' }, - new():: apiVersion + kind, - // Revision indicates the revision of the state represented by Data. - withRevision(revision):: self + { revision: revision }, - mixin:: { - // Data is the serialized representation of the state. - data:: { - local __dataMixin(data) = { data+: data }, - mixinInstance(data):: __dataMixin(data), - // Raw is the underlying serialization of this object. - withRaw(raw):: self + __dataMixin({ Raw: raw }), - }, - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // DEPRECATED - This group version of Deployment is deprecated by apps/v1beta2/Deployment. See the release notes for more information. Deployment enables declarative updates for Pods and ReplicaSets. - deployment:: { - local kind = { kind: 'Deployment' }, - new(name='', replicas=1, containers='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - // Standard object metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the Deployment. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Indicates that the deployment is paused. - withPaused(paused):: self + __specMixin({ paused: paused }), - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + __specMixin({ progressDeadlineSeconds: progressDeadlineSeconds }), - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 2. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // DEPRECATED. The config this deployment is rolling back to. Will be cleared after rollback is done. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = __specMixin({ rollbackTo+: rollbackTo }), - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.apps.v1beta1.rollbackConfig, - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = __specMixin({ strategy+: strategy }), - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1beta1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.apps.v1beta1.deploymentSpec, - }, - }, - // DEPRECATED. DeploymentRollback stores the information required to rollback a deployment. - deploymentRollback:: { - local kind = { kind: 'DeploymentRollback' }, - new(name=''):: apiVersion + kind + self.withName(name), - // Required: This must match the Name of a deployment. - withName(name):: self + { name: name }, - // The annotations to be updated to a deployment - withUpdatedAnnotations(updatedAnnotations):: self + { updatedAnnotations: updatedAnnotations }, - // The annotations to be updated to a deployment - withUpdatedAnnotationsMixin(updatedAnnotations):: self + { updatedAnnotations+: updatedAnnotations }, - mixin:: { - // The config of this deployment rollback. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = { rollbackTo+: rollbackTo }, - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.apps.v1beta1.rollbackConfig, - }, - }, - // Scale represents a scaling request for a resource. - scale:: { - local kind = { kind: 'Scale' }, - new(replicas=1):: apiVersion + kind + self.mixin.spec.withReplicas(replicas), - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // defines the behavior of the scale. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // desired number of instances for the scaled object. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - }, - specType:: hidden.apps.v1beta1.scaleSpec, - }, - }, - // DEPRECATED - This group version of StatefulSet is deprecated by apps/v1beta2/StatefulSet. See the release notes for more information. StatefulSet represents a set of pods with consistent identities. Identities are defined as: - // - Network: A single stable DNS and hostname. - // - Storage: As many VolumeClaims as requested. - // The StatefulSet guarantees that a given network identity will always map to the same storage identity. - statefulSet:: { - local kind = { kind: 'StatefulSet' }, - new(name='', replicas=1, containers='', volumeClaims='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas).withVolumeClaimTemplates(volumeClaims) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the desired identities of pods in this set. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + __specMixin({ podManagementPolicy: podManagementPolicy }), - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // selector is a label query over pods that should match the replica count. If empty, defaulted to labels on the pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + __specMixin({ serviceName: serviceName }), - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta1.statefulSetUpdateStrategy, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates: [volumeClaimTemplates] }), - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates+: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates+: [volumeClaimTemplates] }), - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - }, - specType:: hidden.apps.v1beta1.statefulSetSpec, - }, - }, - }, - v1beta2:: { - local apiVersion = { apiVersion: 'apps/v1beta2' }, - // DEPRECATED - This group version of ControllerRevision is deprecated by apps/v1/ControllerRevision. See the release notes for more information. ControllerRevision implements an immutable snapshot of state data. Clients are responsible for serializing and deserializing the objects that contain their internal state. Once a ControllerRevision has been successfully created, it can not be updated. The API Server will fail validation of all requests that attempt to mutate the Data field. ControllerRevisions may, however, be deleted. Note that, due to its use by both the DaemonSet and StatefulSet controllers for update and rollback, this object is beta. However, it may be subject to name and representation changes in future releases, and clients should not depend on its stability. It is primarily for internal use by controllers. - controllerRevision:: { - local kind = { kind: 'ControllerRevision' }, - new():: apiVersion + kind, - // Revision indicates the revision of the state represented by Data. - withRevision(revision):: self + { revision: revision }, - mixin:: { - // Data is the serialized representation of the state. - data:: { - local __dataMixin(data) = { data+: data }, - mixinInstance(data):: __dataMixin(data), - // Raw is the underlying serialization of this object. - withRaw(raw):: self + __dataMixin({ Raw: raw }), - }, - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // DEPRECATED - This group version of DaemonSet is deprecated by apps/v1/DaemonSet. See the release notes for more information. DaemonSet represents the configuration of a daemon set. - daemonSet:: { - local kind = { kind: 'DaemonSet' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The desired behavior of this daemon set. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta2.daemonSetUpdateStrategy, - }, - specType:: hidden.apps.v1beta2.daemonSetSpec, - }, - }, - // DEPRECATED - This group version of Deployment is deprecated by apps/v1/Deployment. See the release notes for more information. Deployment enables declarative updates for Pods and ReplicaSets. - deployment:: { - local kind = { kind: 'Deployment' }, - new(name='', replicas=1, containers='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - // Standard object metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the Deployment. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Indicates that the deployment is paused. - withPaused(paused):: self + __specMixin({ paused: paused }), - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + __specMixin({ progressDeadlineSeconds: progressDeadlineSeconds }), - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. It must match the pod template's labels. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = __specMixin({ strategy+: strategy }), - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1beta2.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.apps.v1beta2.deploymentSpec, - }, - }, - // DEPRECATED - This group version of ReplicaSet is deprecated by apps/v1/ReplicaSet. See the release notes for more information. ReplicaSet ensures that a specified number of pod replicas are running at any given time. - replicaSet:: { - local kind = { kind: 'ReplicaSet' }, - new():: apiVersion + kind, - mixin:: { - // If the Labels of a ReplicaSet are empty, they are defaulted to be the same as the Pod(s) that the ReplicaSet manages. Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the specification of the desired behavior of the ReplicaSet. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // Selector is a label query over pods that should match the replica count. Label keys and values that must match in order to be controlled by this replica set. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.apps.v1beta2.replicaSetSpec, - }, - }, - // Scale represents a scaling request for a resource. - scale:: { - local kind = { kind: 'Scale' }, - new(replicas=1):: apiVersion + kind + self.mixin.spec.withReplicas(replicas), - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // defines the behavior of the scale. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // desired number of instances for the scaled object. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - }, - specType:: hidden.apps.v1beta2.scaleSpec, - }, - }, - // DEPRECATED - This group version of StatefulSet is deprecated by apps/v1/StatefulSet. See the release notes for more information. StatefulSet represents a set of pods with consistent identities. Identities are defined as: - // - Network: A single stable DNS and hostname. - // - Storage: As many VolumeClaims as requested. - // The StatefulSet guarantees that a given network identity will always map to the same storage identity. - statefulSet:: { - local kind = { kind: 'StatefulSet' }, - new(name='', replicas=1, containers='', volumeClaims='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas).withVolumeClaimTemplates(volumeClaims) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the desired identities of pods in this set. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + __specMixin({ podManagementPolicy: podManagementPolicy }), - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // selector is a label query over pods that should match the replica count. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + __specMixin({ serviceName: serviceName }), - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta2.statefulSetUpdateStrategy, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates: [volumeClaimTemplates] }), - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then __specMixin({ volumeClaimTemplates+: volumeClaimTemplates }) else __specMixin({ volumeClaimTemplates+: [volumeClaimTemplates] }), - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - }, - specType:: hidden.apps.v1beta2.statefulSetSpec, - }, - }, - }, - }, - authentication:: { - v1:: { - local apiVersion = { apiVersion: 'authentication.k8s.io/v1' }, - // TokenReview attempts to authenticate a token to a known user. Note: TokenReview requests may be cached by the webhook token authenticator plugin in the kube-apiserver. - tokenReview:: { - local kind = { kind: 'TokenReview' }, - new(token=''):: apiVersion + kind + self.mixin.spec.withToken(token), - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then __specMixin({ audiences: audiences }) else __specMixin({ audiences: [audiences] }), - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then __specMixin({ audiences+: audiences }) else __specMixin({ audiences+: [audiences] }), - // Token is the opaque bearer token. - withToken(token):: self + __specMixin({ token: token }), - }, - specType:: hidden.authentication.v1.tokenReviewSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'authentication.k8s.io/v1beta1' }, - // TokenReview attempts to authenticate a token to a known user. Note: TokenReview requests may be cached by the webhook token authenticator plugin in the kube-apiserver. - tokenReview:: { - local kind = { kind: 'TokenReview' }, - new(token=''):: apiVersion + kind + self.mixin.spec.withToken(token), - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then __specMixin({ audiences: audiences }) else __specMixin({ audiences: [audiences] }), - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then __specMixin({ audiences+: audiences }) else __specMixin({ audiences+: [audiences] }), - // Token is the opaque bearer token. - withToken(token):: self + __specMixin({ token: token }), - }, - specType:: hidden.authentication.v1beta1.tokenReviewSpec, - }, - }, - }, - }, - authorization:: { - v1:: { - local apiVersion = { apiVersion: 'authorization.k8s.io/v1' }, - // LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. Having a namespace scoped resource makes it much easier to grant namespace scoped policy that includes permissions checking. - localSubjectAccessReview:: { - local kind = { kind: 'LocalSubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. spec.namespace must be equal to the namespace you made the request against. If empty, it is defaulted. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + __specMixin({ extra: extra }), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + __specMixin({ extra+: extra }), - // Groups is the groups you're testing for. - withGroups(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups: groups }) else __specMixin({ groups: [groups] }), - // Groups is the groups you're testing for. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups+: groups }) else __specMixin({ groups+: [groups] }), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1.resourceAttributes, - // UID information about the requesting user. - withUid(uid):: self + __specMixin({ uid: uid }), - // User is the user you're testing for. If you specify "User" but not "Groups", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + __specMixin({ user: user }), - }, - specType:: hidden.authorization.v1.subjectAccessReviewSpec, - }, - }, - // SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a spec.namespace means "in all namespaces". Self is a special case, because users should always be able to check whether they can perform an action - selfSubjectAccessReview:: { - local kind = { kind: 'SelfSubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. user and groups must be empty - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1.resourceAttributes, - }, - specType:: hidden.authorization.v1.selfSubjectAccessReviewSpec, - }, - }, - // SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. The returned list of actions may be incomplete depending on the server's authorization mode, and any errors experienced during the evaluation. SelfSubjectRulesReview should be used by UIs to show/hide actions, or to quickly let an end user reason about their permissions. It should NOT Be used by external systems to drive authorization decisions as this raises confused deputy, cache lifetime/revocation, and correctness concerns. SubjectAccessReview, and LocalAccessReview are the correct way to defer authorization decisions to the API server. - selfSubjectRulesReview:: { - local kind = { kind: 'SelfSubjectRulesReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Namespace to evaluate rules for. Required. - withNamespace(namespace):: self + __specMixin({ namespace: namespace }), - }, - specType:: hidden.authorization.v1.selfSubjectRulesReviewSpec, - }, - }, - // SubjectAccessReview checks whether or not a user or group can perform an action. - subjectAccessReview:: { - local kind = { kind: 'SubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + __specMixin({ extra: extra }), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + __specMixin({ extra+: extra }), - // Groups is the groups you're testing for. - withGroups(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups: groups }) else __specMixin({ groups: [groups] }), - // Groups is the groups you're testing for. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups+: groups }) else __specMixin({ groups+: [groups] }), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1.resourceAttributes, - // UID information about the requesting user. - withUid(uid):: self + __specMixin({ uid: uid }), - // User is the user you're testing for. If you specify "User" but not "Groups", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + __specMixin({ user: user }), - }, - specType:: hidden.authorization.v1.subjectAccessReviewSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'authorization.k8s.io/v1beta1' }, - // LocalSubjectAccessReview checks whether or not a user or group can perform an action in a given namespace. Having a namespace scoped resource makes it much easier to grant namespace scoped policy that includes permissions checking. - localSubjectAccessReview:: { - local kind = { kind: 'LocalSubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. spec.namespace must be equal to the namespace you made the request against. If empty, it is defaulted. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + __specMixin({ extra: extra }), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + __specMixin({ extra+: extra }), - // Groups is the groups you're testing for. - withGroup(group):: self + if std.type(group) == 'array' then __specMixin({ group: group }) else __specMixin({ group: [group] }), - // Groups is the groups you're testing for. - withGroupMixin(group):: self + if std.type(group) == 'array' then __specMixin({ group+: group }) else __specMixin({ group+: [group] }), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1beta1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1beta1.resourceAttributes, - // UID information about the requesting user. - withUid(uid):: self + __specMixin({ uid: uid }), - // User is the user you're testing for. If you specify "User" but not "Group", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + __specMixin({ user: user }), - }, - specType:: hidden.authorization.v1beta1.subjectAccessReviewSpec, - }, - }, - // SelfSubjectAccessReview checks whether or the current user can perform an action. Not filling in a spec.namespace means "in all namespaces". Self is a special case, because users should always be able to check whether they can perform an action - selfSubjectAccessReview:: { - local kind = { kind: 'SelfSubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. user and groups must be empty - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1beta1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1beta1.resourceAttributes, - }, - specType:: hidden.authorization.v1beta1.selfSubjectAccessReviewSpec, - }, - }, - // SelfSubjectRulesReview enumerates the set of actions the current user can perform within a namespace. The returned list of actions may be incomplete depending on the server's authorization mode, and any errors experienced during the evaluation. SelfSubjectRulesReview should be used by UIs to show/hide actions, or to quickly let an end user reason about their permissions. It should NOT Be used by external systems to drive authorization decisions as this raises confused deputy, cache lifetime/revocation, and correctness concerns. SubjectAccessReview, and LocalAccessReview are the correct way to defer authorization decisions to the API server. - selfSubjectRulesReview:: { - local kind = { kind: 'SelfSubjectRulesReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Namespace to evaluate rules for. Required. - withNamespace(namespace):: self + __specMixin({ namespace: namespace }), - }, - specType:: hidden.authorization.v1beta1.selfSubjectRulesReviewSpec, - }, - }, - // SubjectAccessReview checks whether or not a user or group can perform an action. - subjectAccessReview:: { - local kind = { kind: 'SubjectAccessReview' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec holds information about the request being evaluated - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + __specMixin({ extra: extra }), - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + __specMixin({ extra+: extra }), - // Groups is the groups you're testing for. - withGroup(group):: self + if std.type(group) == 'array' then __specMixin({ group: group }) else __specMixin({ group: [group] }), - // Groups is the groups you're testing for. - withGroupMixin(group):: self + if std.type(group) == 'array' then __specMixin({ group+: group }) else __specMixin({ group+: [group] }), - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = __specMixin({ nonResourceAttributes+: nonResourceAttributes }), - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1beta1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = __specMixin({ resourceAttributes+: resourceAttributes }), - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1beta1.resourceAttributes, - // UID information about the requesting user. - withUid(uid):: self + __specMixin({ uid: uid }), - // User is the user you're testing for. If you specify "User" but not "Group", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + __specMixin({ user: user }), - }, - specType:: hidden.authorization.v1beta1.subjectAccessReviewSpec, - }, - }, - }, - }, - autoscaling:: { - v1:: { - local apiVersion = { apiVersion: 'autoscaling/v1' }, - // configuration of a horizontal pod autoscaler. - horizontalPodAutoscaler:: { - local kind = { kind: 'HorizontalPodAutoscaler' }, - new():: apiVersion + kind, - mixin:: { - // Standard object metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // behaviour of autoscaler. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // upper limit for the number of pods that can be set by the autoscaler; cannot be smaller than MinReplicas. - withMaxReplicas(maxReplicas):: self + __specMixin({ maxReplicas: maxReplicas }), - // lower limit for the number of pods that can be set by the autoscaler, default 1. - withMinReplicas(minReplicas):: self + __specMixin({ minReplicas: minReplicas }), - // reference to scaled resource; horizontal pod autoscaler will learn the current resource consumption and will set the desired number of pods by using its Scale subresource. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = __specMixin({ scaleTargetRef+: scaleTargetRef }), - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v1.crossVersionObjectReference, - // target average CPU utilization (represented as a percentage of requested CPU) over all the pods; if not specified the default autoscaling policy will be used. - withTargetCpuUtilizationPercentage(targetCpuUtilizationPercentage):: self + __specMixin({ targetCPUUtilizationPercentage: targetCpuUtilizationPercentage }), - }, - specType:: hidden.autoscaling.v1.horizontalPodAutoscalerSpec, - }, - }, - // Scale represents a scaling request for a resource. - scale:: { - local kind = { kind: 'Scale' }, - new(replicas=1):: apiVersion + kind + self.mixin.spec.withReplicas(replicas), - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // defines the behavior of the scale. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // desired number of instances for the scaled object. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - }, - specType:: hidden.autoscaling.v1.scaleSpec, - }, - }, - }, - v2beta1:: { - local apiVersion = { apiVersion: 'autoscaling/v2beta1' }, - // HorizontalPodAutoscaler is the configuration for a horizontal pod autoscaler, which automatically manages the replica count of any resource implementing the scale subresource based on the metrics specified. - horizontalPodAutoscaler:: { - local kind = { kind: 'HorizontalPodAutoscaler' }, - new():: apiVersion + kind, - mixin:: { - // metadata is the standard object metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // spec is the specification for the behaviour of the autoscaler. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas. - withMaxReplicas(maxReplicas):: self + __specMixin({ maxReplicas: maxReplicas }), - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. - withMetrics(metrics):: self + if std.type(metrics) == 'array' then __specMixin({ metrics: metrics }) else __specMixin({ metrics: [metrics] }), - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. - withMetricsMixin(metrics):: self + if std.type(metrics) == 'array' then __specMixin({ metrics+: metrics }) else __specMixin({ metrics+: [metrics] }), - metricsType:: hidden.autoscaling.v2beta1.metricSpec, - // minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. - withMinReplicas(minReplicas):: self + __specMixin({ minReplicas: minReplicas }), - // scaleTargetRef points to the target resource to scale, and is used to the pods for which metrics should be collected, as well as to actually change the replica count. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = __specMixin({ scaleTargetRef+: scaleTargetRef }), - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - }, - specType:: hidden.autoscaling.v2beta1.horizontalPodAutoscalerSpec, - }, - }, - }, - v2beta2:: { - local apiVersion = { apiVersion: 'autoscaling/v2beta2' }, - // HorizontalPodAutoscaler is the configuration for a horizontal pod autoscaler, which automatically manages the replica count of any resource implementing the scale subresource based on the metrics specified. - horizontalPodAutoscaler:: { - local kind = { kind: 'HorizontalPodAutoscaler' }, - new():: apiVersion + kind, - mixin:: { - // metadata is the standard object metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // spec is the specification for the behaviour of the autoscaler. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas. - withMaxReplicas(maxReplicas):: self + __specMixin({ maxReplicas: maxReplicas }), - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization. - withMetrics(metrics):: self + if std.type(metrics) == 'array' then __specMixin({ metrics: metrics }) else __specMixin({ metrics: [metrics] }), - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization. - withMetricsMixin(metrics):: self + if std.type(metrics) == 'array' then __specMixin({ metrics+: metrics }) else __specMixin({ metrics+: [metrics] }), - metricsType:: hidden.autoscaling.v2beta2.metricSpec, - // minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. - withMinReplicas(minReplicas):: self + __specMixin({ minReplicas: minReplicas }), - // scaleTargetRef points to the target resource to scale, and is used to the pods for which metrics should be collected, as well as to actually change the replica count. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = __specMixin({ scaleTargetRef+: scaleTargetRef }), - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - }, - specType:: hidden.autoscaling.v2beta2.horizontalPodAutoscalerSpec, - }, - }, - }, - }, - batch:: { - v1:: { - local apiVersion = { apiVersion: 'batch/v1' }, - // Job represents the configuration of a single job. - job:: { - local kind = { kind: 'Job' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of a job. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // Specifies the number of retries before marking this job failed. Defaults to 6 - withBackoffLimit(backoffLimit):: self + __specMixin({ backoffLimit: backoffLimit }), - // Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withCompletions(completions):: self + __specMixin({ completions: completions }), - // manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector - withManualSelector(manualSelector):: self + __specMixin({ manualSelector: manualSelector }), - // Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withParallelism(parallelism):: self + __specMixin({ parallelism: parallelism }), - // A label query over pods that should match the pod count. Normally, the system sets this field for you. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Describes the pod that will be created when executing a job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature. - withTtlSecondsAfterFinished(ttlSecondsAfterFinished):: self + __specMixin({ ttlSecondsAfterFinished: ttlSecondsAfterFinished }), - }, - specType:: hidden.batch.v1.jobSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'batch/v1beta1' }, - // CronJob represents the configuration of a single cron job. - cronJob:: { - local kind = { kind: 'CronJob' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of a cron job, including the schedule. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Specifies how to treat concurrent executions of a Job. Valid values are: - "Allow" (default): allows CronJobs to run concurrently; - "Forbid": forbids concurrent runs, skipping next run if previous run hasn't finished yet; - "Replace": cancels currently running job and replaces it with a new one - withConcurrencyPolicy(concurrencyPolicy):: self + __specMixin({ concurrencyPolicy: concurrencyPolicy }), - // The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withFailedJobsHistoryLimit(failedJobsHistoryLimit):: self + __specMixin({ failedJobsHistoryLimit: failedJobsHistoryLimit }), - // Specifies the job that will be created when executing a CronJob. - jobTemplate:: { - local __jobTemplateMixin(jobTemplate) = __specMixin({ jobTemplate+: jobTemplate }), - mixinInstance(jobTemplate):: __jobTemplateMixin(jobTemplate), - // Standard object's metadata of the jobs created from this template. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __jobTemplateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the job. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __jobTemplateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // Specifies the number of retries before marking this job failed. Defaults to 6 - withBackoffLimit(backoffLimit):: self + __specMixin({ backoffLimit: backoffLimit }), - // Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withCompletions(completions):: self + __specMixin({ completions: completions }), - // manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector - withManualSelector(manualSelector):: self + __specMixin({ manualSelector: manualSelector }), - // Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withParallelism(parallelism):: self + __specMixin({ parallelism: parallelism }), - // A label query over pods that should match the pod count. Normally, the system sets this field for you. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Describes the pod that will be created when executing a job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature. - withTtlSecondsAfterFinished(ttlSecondsAfterFinished):: self + __specMixin({ ttlSecondsAfterFinished: ttlSecondsAfterFinished }), - }, - specType:: hidden.batch.v1.jobSpec, - }, - jobTemplateType:: hidden.batch.v1beta1.jobTemplateSpec, - // The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron. - withSchedule(schedule):: self + __specMixin({ schedule: schedule }), - // Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones. - withStartingDeadlineSeconds(startingDeadlineSeconds):: self + __specMixin({ startingDeadlineSeconds: startingDeadlineSeconds }), - // The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 3. - withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit):: self + __specMixin({ successfulJobsHistoryLimit: successfulJobsHistoryLimit }), - // This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false. - withSuspend(suspend):: self + __specMixin({ suspend: suspend }), - }, - specType:: hidden.batch.v1beta1.cronJobSpec, - }, - }, - }, - }, - certificates:: { - v1beta1:: { - local apiVersion = { apiVersion: 'certificates.k8s.io/v1beta1' }, - // Describes a certificate signing request - certificateSigningRequest:: { - local kind = { kind: 'CertificateSigningRequest' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The certificate request itself and any additional information. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Extra information about the requesting user. See user.Info interface for details. - withExtra(extra):: self + __specMixin({ extra: extra }), - // Extra information about the requesting user. See user.Info interface for details. - withExtraMixin(extra):: self + __specMixin({ extra+: extra }), - // Group information about the requesting user. See user.Info interface for details. - withGroups(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups: groups }) else __specMixin({ groups: [groups] }), - // Group information about the requesting user. See user.Info interface for details. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then __specMixin({ groups+: groups }) else __specMixin({ groups+: [groups] }), - // Base64-encoded PKCS#10 CSR data - withRequest(request):: self + __specMixin({ request: request }), - // UID information about the requesting user. See user.Info interface for details. - withUid(uid):: self + __specMixin({ uid: uid }), - // allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 - // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 - withUsages(usages):: self + if std.type(usages) == 'array' then __specMixin({ usages: usages }) else __specMixin({ usages: [usages] }), - // allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 - // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 - withUsagesMixin(usages):: self + if std.type(usages) == 'array' then __specMixin({ usages+: usages }) else __specMixin({ usages+: [usages] }), - // Information about the requesting user. See user.Info interface for details. - withUsername(username):: self + __specMixin({ username: username }), - }, - specType:: hidden.certificates.v1beta1.certificateSigningRequestSpec, - }, - }, - }, - }, - coordination:: { - v1:: { - local apiVersion = { apiVersion: 'coordination.k8s.io/v1' }, - // Lease defines a lease concept. - lease:: { - local kind = { kind: 'Lease' }, - new():: apiVersion + kind, - mixin:: { - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the Lease. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // acquireTime is a time when the current lease was acquired. - withAcquireTime(acquireTime):: self + __specMixin({ acquireTime: acquireTime }), - // holderIdentity contains the identity of the holder of a current lease. - withHolderIdentity(holderIdentity):: self + __specMixin({ holderIdentity: holderIdentity }), - // leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime. - withLeaseDurationSeconds(leaseDurationSeconds):: self + __specMixin({ leaseDurationSeconds: leaseDurationSeconds }), - // leaseTransitions is the number of transitions of a lease between holders. - withLeaseTransitions(leaseTransitions):: self + __specMixin({ leaseTransitions: leaseTransitions }), - // renewTime is a time when the current holder of a lease has last updated the lease. - withRenewTime(renewTime):: self + __specMixin({ renewTime: renewTime }), - }, - specType:: hidden.coordination.v1.leaseSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'coordination.k8s.io/v1beta1' }, - // Lease defines a lease concept. - lease:: { - local kind = { kind: 'Lease' }, - new():: apiVersion + kind, - mixin:: { - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the Lease. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // acquireTime is a time when the current lease was acquired. - withAcquireTime(acquireTime):: self + __specMixin({ acquireTime: acquireTime }), - // holderIdentity contains the identity of the holder of a current lease. - withHolderIdentity(holderIdentity):: self + __specMixin({ holderIdentity: holderIdentity }), - // leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime. - withLeaseDurationSeconds(leaseDurationSeconds):: self + __specMixin({ leaseDurationSeconds: leaseDurationSeconds }), - // leaseTransitions is the number of transitions of a lease between holders. - withLeaseTransitions(leaseTransitions):: self + __specMixin({ leaseTransitions: leaseTransitions }), - // renewTime is a time when the current holder of a lease has last updated the lease. - withRenewTime(renewTime):: self + __specMixin({ renewTime: renewTime }), - }, - specType:: hidden.coordination.v1beta1.leaseSpec, - }, - }, - }, - }, - core:: { - v1:: { - local apiVersion = { apiVersion: 'v1' }, - // Binding ties one object to another; for example, a pod is bound to a node by a scheduler. Deprecated in 1.7, please use the bindings subresource of pods instead. - binding:: { - local kind = { kind: 'Binding' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The target object that you want to bind to the standard object. - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __targetMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __targetMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __targetMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __targetMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __targetMixin({ uid: uid }), - }, - targetType:: hidden.core.v1.objectReference, - }, - }, - // ConfigMap holds configuration data for pods to consume. - configMap:: { - local kind = { kind: 'ConfigMap' }, - new(name='', data=''):: apiVersion + kind + self.withData(data) + self.mixin.metadata.withName(name), - // BinaryData contains the binary data. Each key must consist of alphanumeric characters, '-', '_' or '.'. BinaryData can contain byte sequences that are not in the UTF-8 range. The keys stored in BinaryData must not overlap with the ones in the Data field, this is enforced during validation process. Using this field will require 1.10+ apiserver and kubelet. - withBinaryData(binaryData):: self + { binaryData: binaryData }, - // BinaryData contains the binary data. Each key must consist of alphanumeric characters, '-', '_' or '.'. BinaryData can contain byte sequences that are not in the UTF-8 range. The keys stored in BinaryData must not overlap with the ones in the Data field, this is enforced during validation process. Using this field will require 1.10+ apiserver and kubelet. - withBinaryDataMixin(binaryData):: self + { binaryData+: binaryData }, - // Data contains the configuration data. Each key must consist of alphanumeric characters, '-', '_' or '.'. Values with non-UTF-8 byte sequences must use the BinaryData field. The keys stored in Data must not overlap with the keys in the BinaryData field, this is enforced during validation process. - withData(data):: self + { data: data }, - // Data contains the configuration data. Each key must consist of alphanumeric characters, '-', '_' or '.'. Values with non-UTF-8 byte sequences must use the BinaryData field. The keys stored in Data must not overlap with the keys in the BinaryData field, this is enforced during validation process. - withDataMixin(data):: self + { data+: data }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // Endpoints is a collection of endpoints that implement the actual service. Example: - // Name: "mysvc", - // Subsets: [ - // { - // Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}], - // Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}] - // }, - // { - // Addresses: [{"ip": "10.10.3.3"}], - // Ports: [{"name": "a", "port": 93}, {"name": "b", "port": 76}] - // }, - // ] - endpoints:: { - local kind = { kind: 'Endpoints' }, - new():: apiVersion + kind, - // The set of all endpoints is the union of all subsets. Addresses are placed into subsets according to the IPs they share. A single address with multiple ports, some of which are ready and some of which are not (because they come from different containers) will result in the address being displayed in different subsets for the different ports. No address will appear in both Addresses and NotReadyAddresses in the same subset. Sets of addresses and ports that comprise a service. - withSubsets(subsets):: self + if std.type(subsets) == 'array' then { subsets: subsets } else { subsets: [subsets] }, - // The set of all endpoints is the union of all subsets. Addresses are placed into subsets according to the IPs they share. A single address with multiple ports, some of which are ready and some of which are not (because they come from different containers) will result in the address being displayed in different subsets for the different ports. No address will appear in both Addresses and NotReadyAddresses in the same subset. Sets of addresses and ports that comprise a service. - withSubsetsMixin(subsets):: self + if std.type(subsets) == 'array' then { subsets+: subsets } else { subsets+: [subsets] }, - subsetsType:: hidden.core.v1.endpointSubset, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // Event is a report of an event somewhere in the cluster. - event:: { - local kind = { kind: 'Event' }, - new():: apiVersion + kind, - // What action was taken/failed regarding to the Regarding object. - withAction(action):: self + { action: action }, - // The number of times this event has occurred. - withCount(count):: self + { count: count }, - // Time when this Event was first observed. - withEventTime(eventTime):: self + { eventTime: eventTime }, - // The time at which the event was first recorded. (Time of server receipt is in TypeMeta.) - withFirstTimestamp(firstTimestamp):: self + { firstTimestamp: firstTimestamp }, - // The time at which the most recent occurrence of this event was recorded. - withLastTimestamp(lastTimestamp):: self + { lastTimestamp: lastTimestamp }, - // A human-readable description of the status of this operation. - withMessage(message):: self + { message: message }, - // This should be a short, machine understandable string that gives the reason for the transition into the object's current status. - withReason(reason):: self + { reason: reason }, - // Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. - withReportingComponent(reportingComponent):: self + { reportingComponent: reportingComponent }, - // ID of the controller instance, e.g. `kubelet-xyzf`. - withReportingInstance(reportingInstance):: self + { reportingInstance: reportingInstance }, - // Type of this event (Normal, Warning), new types could be added in the future - withType(type):: self + { type: type }, - mixin:: { - // The object that this event is about. - involvedObject:: { - local __involvedObjectMixin(involvedObject) = { involvedObject+: involvedObject }, - mixinInstance(involvedObject):: __involvedObjectMixin(involvedObject), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __involvedObjectMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __involvedObjectMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __involvedObjectMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __involvedObjectMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __involvedObjectMixin({ uid: uid }), - }, - involvedObjectType:: hidden.core.v1.objectReference, - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Optional secondary object for more complex actions. - related:: { - local __relatedMixin(related) = { related+: related }, - mixinInstance(related):: __relatedMixin(related), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __relatedMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __relatedMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __relatedMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __relatedMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __relatedMixin({ uid: uid }), - }, - relatedType:: hidden.core.v1.objectReference, - // Data about the Event series this event represents or nil if it's a singleton Event. - series:: { - local __seriesMixin(series) = { series+: series }, - mixinInstance(series):: __seriesMixin(series), - // Number of occurrences in this series up to the last heartbeat time - withCount(count):: self + __seriesMixin({ count: count }), - // Time of the last occurrence observed - withLastObservedTime(lastObservedTime):: self + __seriesMixin({ lastObservedTime: lastObservedTime }), - // State of this Series: Ongoing or Finished - withState(state):: self + __seriesMixin({ state: state }), - }, - seriesType:: hidden.core.v1.eventSeries, - // The component reporting this event. Should be a short machine understandable string. - source:: { - local __sourceMixin(source) = { source+: source }, - mixinInstance(source):: __sourceMixin(source), - // Component from which the event is generated. - withComponent(component):: self + __sourceMixin({ component: component }), - // Node name on which the event is generated. - withHost(host):: self + __sourceMixin({ host: host }), - }, - sourceType:: hidden.core.v1.eventSource, - }, - }, - // LimitRange sets resource usage limits for each kind of resource in a Namespace. - limitRange:: { - local kind = { kind: 'LimitRange' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the limits enforced. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Limits is the list of LimitRangeItem objects that are enforced. - withLimits(limits):: self + if std.type(limits) == 'array' then __specMixin({ limits: limits }) else __specMixin({ limits: [limits] }), - // Limits is the list of LimitRangeItem objects that are enforced. - withLimitsMixin(limits):: self + if std.type(limits) == 'array' then __specMixin({ limits+: limits }) else __specMixin({ limits+: [limits] }), - limitsType:: hidden.core.v1.limitRangeItem, - }, - specType:: hidden.core.v1.limitRangeSpec, - }, - }, - // Namespace provides a scope for Names. Use of multiple namespaces is optional. - namespace:: { - local kind = { kind: 'Namespace' }, - new(name=''):: apiVersion + kind + self.mixin.metadata.withName(name), - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the behavior of the Namespace. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/ - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __specMixin({ finalizers: finalizers }) else __specMixin({ finalizers: [finalizers] }), - // Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/ - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __specMixin({ finalizers+: finalizers }) else __specMixin({ finalizers+: [finalizers] }), - }, - specType:: hidden.core.v1.namespaceSpec, - }, - }, - // Node is a worker node in Kubernetes. Each node will have a unique identifier in the cache (i.e. in etcd). - node:: { - local kind = { kind: 'Node' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the behavior of a node. https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // If specified, the source to get node configuration from The DynamicKubeletConfig feature gate must be enabled for the Kubelet to use this field - configSource:: { - local __configSourceMixin(configSource) = __specMixin({ configSource+: configSource }), - mixinInstance(configSource):: __configSourceMixin(configSource), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __configSourceMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - configSourceType:: hidden.core.v1.nodeConfigSource, - // Deprecated. Not all kubelets will set this field. Remove field after 1.13. see: https://issues.k8s.io/61966 - withExternalId(externalId):: self + __specMixin({ externalID: externalId }), - // PodCIDR represents the pod IP range assigned to the node. - withPodCidr(podCidr):: self + __specMixin({ podCIDR: podCidr }), - // ID of the node assigned by the cloud provider in the format: :// - withProviderId(providerId):: self + __specMixin({ providerID: providerId }), - // If specified, the node's taints. - withTaints(taints):: self + if std.type(taints) == 'array' then __specMixin({ taints: taints }) else __specMixin({ taints: [taints] }), - // If specified, the node's taints. - withTaintsMixin(taints):: self + if std.type(taints) == 'array' then __specMixin({ taints+: taints }) else __specMixin({ taints+: [taints] }), - taintsType:: hidden.core.v1.taint, - // Unschedulable controls node schedulability of new pods. By default, node is schedulable. More info: https://kubernetes.io/docs/concepts/nodes/node/#manual-node-administration - withUnschedulable(unschedulable):: self + __specMixin({ unschedulable: unschedulable }), - }, - specType:: hidden.core.v1.nodeSpec, - }, - }, - // PersistentVolume (PV) is a storage resource provisioned by an administrator. It is analogous to a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes - persistentVolume:: { - local kind = { kind: 'PersistentVolume' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines a specification of a persistent volume owned by the cluster. Provisioned by an administrator. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistent-volumes - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes - withAccessModes(accessModes):: self + if std.type(accessModes) == 'array' then __specMixin({ accessModes: accessModes }) else __specMixin({ accessModes: [accessModes] }), - // AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes - withAccessModesMixin(accessModes):: self + if std.type(accessModes) == 'array' then __specMixin({ accessModes+: accessModes }) else __specMixin({ accessModes+: [accessModes] }), - // AWSElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - awsElasticBlockStore:: { - local __awsElasticBlockStoreMixin(awsElasticBlockStore) = __specMixin({ awsElasticBlockStore+: awsElasticBlockStore }), - mixinInstance(awsElasticBlockStore):: __awsElasticBlockStoreMixin(awsElasticBlockStore), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withFsType(fsType):: self + __awsElasticBlockStoreMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). - withPartition(partition):: self + __awsElasticBlockStoreMixin({ partition: partition }), - // Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withReadOnly(readOnly):: self + __awsElasticBlockStoreMixin({ readOnly: readOnly }), - // Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withVolumeId(volumeId):: self + __awsElasticBlockStoreMixin({ volumeID: volumeId }), - }, - awsElasticBlockStoreType:: hidden.core.v1.awsElasticBlockStoreVolumeSource, - // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod. - azureDisk:: { - local __azureDiskMixin(azureDisk) = __specMixin({ azureDisk+: azureDisk }), - mixinInstance(azureDisk):: __azureDiskMixin(azureDisk), - // Host Caching mode: None, Read Only, Read Write. - withCachingMode(cachingMode):: self + __azureDiskMixin({ cachingMode: cachingMode }), - // The Name of the data disk in the blob storage - withDiskName(diskName):: self + __azureDiskMixin({ diskName: diskName }), - // The URI the data disk in the blob storage - withDiskUri(diskUri):: self + __azureDiskMixin({ diskURI: diskUri }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __azureDiskMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureDiskMixin({ readOnly: readOnly }), - }, - azureDiskType:: hidden.core.v1.azureDiskVolumeSource, - // AzureFile represents an Azure File Service mount on the host and bind mount to the pod. - azureFile:: { - local __azureFileMixin(azureFile) = __specMixin({ azureFile+: azureFile }), - mixinInstance(azureFile):: __azureFileMixin(azureFile), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureFileMixin({ readOnly: readOnly }), - // the name of secret that contains Azure Storage Account Name and Key - withSecretName(secretName):: self + __azureFileMixin({ secretName: secretName }), - // the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod - withSecretNamespace(secretNamespace):: self + __azureFileMixin({ secretNamespace: secretNamespace }), - // Share Name - withShareName(shareName):: self + __azureFileMixin({ shareName: shareName }), - }, - azureFileType:: hidden.core.v1.azureFilePersistentVolumeSource, - // A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacity(capacity):: self + __specMixin({ capacity: capacity }), - // A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacityMixin(capacity):: self + __specMixin({ capacity+: capacity }), - // CephFS represents a Ceph FS mount on the host that shares a pod's lifetime - cephfs:: { - local __cephfsMixin(cephfs) = __specMixin({ cephfs+: cephfs }), - mixinInstance(cephfs):: __cephfsMixin(cephfs), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors: monitors }) else __cephfsMixin({ monitors: [monitors] }), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors+: monitors }) else __cephfsMixin({ monitors+: [monitors] }), - // Optional: Used as the mounted root, rather than the full Ceph tree, default is / - withPath(path):: self + __cephfsMixin({ path: path }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withReadOnly(readOnly):: self + __cephfsMixin({ readOnly: readOnly }), - // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withSecretFile(secretFile):: self + __cephfsMixin({ secretFile: secretFile }), - // Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __cephfsMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // Optional: User is the rados user name, default is admin More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withUser(user):: self + __cephfsMixin({ user: user }), - }, - cephfsType:: hidden.core.v1.cephFsPersistentVolumeSource, - // Cinder represents a cinder volume attached and mounted on kubelets host machine More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - cinder:: { - local __cinderMixin(cinder) = __specMixin({ cinder+: cinder }), - mixinInstance(cinder):: __cinderMixin(cinder), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withFsType(fsType):: self + __cinderMixin({ fsType: fsType }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withReadOnly(readOnly):: self + __cinderMixin({ readOnly: readOnly }), - // Optional: points to a secret object containing parameters used to connect to OpenStack. - secretRef:: { - local __secretRefMixin(secretRef) = __cinderMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // volume id used to identify the volume in cinder More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withVolumeId(volumeId):: self + __cinderMixin({ volumeID: volumeId }), - }, - cinderType:: hidden.core.v1.cinderPersistentVolumeSource, - // ClaimRef is part of a bi-directional binding between PersistentVolume and PersistentVolumeClaim. Expected to be non-nil when bound. claim.VolumeName is the authoritative bind between PV and PVC. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#binding - claimRef:: { - local __claimRefMixin(claimRef) = __specMixin({ claimRef+: claimRef }), - mixinInstance(claimRef):: __claimRefMixin(claimRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __claimRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __claimRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __claimRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __claimRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __claimRefMixin({ uid: uid }), - }, - claimRefType:: hidden.core.v1.objectReference, - // CSI represents storage that is handled by an external CSI driver (Beta feature). - csi:: { - local __csiMixin(csi) = __specMixin({ csi+: csi }), - mixinInstance(csi):: __csiMixin(csi), - // ControllerPublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI ControllerPublishVolume and ControllerUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - controllerPublishSecretRef:: { - local __controllerPublishSecretRefMixin(controllerPublishSecretRef) = __csiMixin({ controllerPublishSecretRef+: controllerPublishSecretRef }), - mixinInstance(controllerPublishSecretRef):: __controllerPublishSecretRefMixin(controllerPublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __controllerPublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __controllerPublishSecretRefMixin({ namespace: namespace }), - }, - controllerPublishSecretRefType:: hidden.core.v1.secretReference, - // Driver is the name of the driver to use for this volume. Required. - withDriver(driver):: self + __csiMixin({ driver: driver }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". - withFsType(fsType):: self + __csiMixin({ fsType: fsType }), - // NodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodePublishSecretRef:: { - local __nodePublishSecretRefMixin(nodePublishSecretRef) = __csiMixin({ nodePublishSecretRef+: nodePublishSecretRef }), - mixinInstance(nodePublishSecretRef):: __nodePublishSecretRefMixin(nodePublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodePublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodePublishSecretRefMixin({ namespace: namespace }), - }, - nodePublishSecretRefType:: hidden.core.v1.secretReference, - // NodeStageSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodeStageVolume and NodeStageVolume and NodeUnstageVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodeStageSecretRef:: { - local __nodeStageSecretRefMixin(nodeStageSecretRef) = __csiMixin({ nodeStageSecretRef+: nodeStageSecretRef }), - mixinInstance(nodeStageSecretRef):: __nodeStageSecretRefMixin(nodeStageSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodeStageSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodeStageSecretRefMixin({ namespace: namespace }), - }, - nodeStageSecretRefType:: hidden.core.v1.secretReference, - // Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write). - withReadOnly(readOnly):: self + __csiMixin({ readOnly: readOnly }), - // Attributes of the volume to publish. - withVolumeAttributes(volumeAttributes):: self + __csiMixin({ volumeAttributes: volumeAttributes }), - // Attributes of the volume to publish. - withVolumeAttributesMixin(volumeAttributes):: self + __csiMixin({ volumeAttributes+: volumeAttributes }), - // VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required. - withVolumeHandle(volumeHandle):: self + __csiMixin({ volumeHandle: volumeHandle }), - }, - csiType:: hidden.core.v1.csiPersistentVolumeSource, - // FC represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod. - fc:: { - local __fcMixin(fc) = __specMixin({ fc+: fc }), - mixinInstance(fc):: __fcMixin(fc), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __fcMixin({ fsType: fsType }), - // Optional: FC target lun number - withLun(lun):: self + __fcMixin({ lun: lun }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __fcMixin({ readOnly: readOnly }), - // Optional: FC target worldwide names (WWNs) - withTargetWwns(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs: targetWwns }) else __fcMixin({ targetWWNs: [targetWwns] }), - // Optional: FC target worldwide names (WWNs) - withTargetWwnsMixin(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs+: targetWwns }) else __fcMixin({ targetWWNs+: [targetWwns] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwids(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids: wwids }) else __fcMixin({ wwids: [wwids] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwidsMixin(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids+: wwids }) else __fcMixin({ wwids+: [wwids] }), - }, - fcType:: hidden.core.v1.fcVolumeSource, - // FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin. - flexVolume:: { - local __flexVolumeMixin(flexVolume) = __specMixin({ flexVolume+: flexVolume }), - mixinInstance(flexVolume):: __flexVolumeMixin(flexVolume), - // Driver is the name of the driver to use for this volume. - withDriver(driver):: self + __flexVolumeMixin({ driver: driver }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script. - withFsType(fsType):: self + __flexVolumeMixin({ fsType: fsType }), - // Optional: Extra command options if any. - withOptions(options):: self + __flexVolumeMixin({ options: options }), - // Optional: Extra command options if any. - withOptionsMixin(options):: self + __flexVolumeMixin({ options+: options }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __flexVolumeMixin({ readOnly: readOnly }), - // Optional: SecretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts. - secretRef:: { - local __secretRefMixin(secretRef) = __flexVolumeMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - flexVolumeType:: hidden.core.v1.flexPersistentVolumeSource, - // Flocker represents a Flocker volume attached to a kubelet's host machine and exposed to the pod for its usage. This depends on the Flocker control service being running - flocker:: { - local __flockerMixin(flocker) = __specMixin({ flocker+: flocker }), - mixinInstance(flocker):: __flockerMixin(flocker), - // Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated - withDatasetName(datasetName):: self + __flockerMixin({ datasetName: datasetName }), - // UUID of the dataset. This is unique identifier of a Flocker dataset - withDatasetUuid(datasetUuid):: self + __flockerMixin({ datasetUUID: datasetUuid }), - }, - flockerType:: hidden.core.v1.flockerVolumeSource, - // GCEPersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. Provisioned by an admin. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - gcePersistentDisk:: { - local __gcePersistentDiskMixin(gcePersistentDisk) = __specMixin({ gcePersistentDisk+: gcePersistentDisk }), - mixinInstance(gcePersistentDisk):: __gcePersistentDiskMixin(gcePersistentDisk), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withFsType(fsType):: self + __gcePersistentDiskMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPartition(partition):: self + __gcePersistentDiskMixin({ partition: partition }), - // Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPdName(pdName):: self + __gcePersistentDiskMixin({ pdName: pdName }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withReadOnly(readOnly):: self + __gcePersistentDiskMixin({ readOnly: readOnly }), - }, - gcePersistentDiskType:: hidden.core.v1.gcePersistentDiskVolumeSource, - // Glusterfs represents a Glusterfs volume that is attached to a host and exposed to the pod. Provisioned by an admin. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md - glusterfs:: { - local __glusterfsMixin(glusterfs) = __specMixin({ glusterfs+: glusterfs }), - mixinInstance(glusterfs):: __glusterfsMixin(glusterfs), - // EndpointsName is the endpoint name that details Glusterfs topology. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpoints(endpoints):: self + __glusterfsMixin({ endpoints: endpoints }), - // EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpointsNamespace(endpointsNamespace):: self + __glusterfsMixin({ endpointsNamespace: endpointsNamespace }), - // Path is the Glusterfs volume path. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withPath(path):: self + __glusterfsMixin({ path: path }), - // ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withReadOnly(readOnly):: self + __glusterfsMixin({ readOnly: readOnly }), - }, - glusterfsType:: hidden.core.v1.glusterfsPersistentVolumeSource, - // HostPath represents a directory on the host. Provisioned by a developer or tester. This is useful for single-node development and testing only! On-host storage is not supported in any way and WILL NOT WORK in a multi-node cluster. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - hostPath:: { - local __hostPathMixin(hostPath) = __specMixin({ hostPath+: hostPath }), - mixinInstance(hostPath):: __hostPathMixin(hostPath), - // Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withPath(path):: self + __hostPathMixin({ path: path }), - // Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withType(type):: self + __hostPathMixin({ type: type }), - }, - hostPathType:: hidden.core.v1.hostPathVolumeSource, - // ISCSI represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. Provisioned by an admin. - iscsi:: { - local __iscsiMixin(iscsi) = __specMixin({ iscsi+: iscsi }), - mixinInstance(iscsi):: __iscsiMixin(iscsi), - // whether support iSCSI Discovery CHAP authentication - withChapAuthDiscovery(chapAuthDiscovery):: self + __iscsiMixin({ chapAuthDiscovery: chapAuthDiscovery }), - // whether support iSCSI Session CHAP authentication - withChapAuthSession(chapAuthSession):: self + __iscsiMixin({ chapAuthSession: chapAuthSession }), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi - withFsType(fsType):: self + __iscsiMixin({ fsType: fsType }), - // Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection. - withInitiatorName(initiatorName):: self + __iscsiMixin({ initiatorName: initiatorName }), - // Target iSCSI Qualified Name. - withIqn(iqn):: self + __iscsiMixin({ iqn: iqn }), - // iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp). - withIscsiInterface(iscsiInterface):: self + __iscsiMixin({ iscsiInterface: iscsiInterface }), - // iSCSI Target Lun number. - withLun(lun):: self + __iscsiMixin({ lun: lun }), - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortals(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals: portals }) else __iscsiMixin({ portals: [portals] }), - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortalsMixin(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals+: portals }) else __iscsiMixin({ portals+: [portals] }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. - withReadOnly(readOnly):: self + __iscsiMixin({ readOnly: readOnly }), - // CHAP Secret for iSCSI target and initiator authentication - secretRef:: { - local __secretRefMixin(secretRef) = __iscsiMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withTargetPortal(targetPortal):: self + __iscsiMixin({ targetPortal: targetPortal }), - }, - iscsiType:: hidden.core.v1.iscsiPersistentVolumeSource, - // Local represents directly-attached storage with node affinity - localStorage:: { - local __localStorageMixin(localStorage) = __specMixin({ localStorage+: localStorage }), - mixinInstance(localStorage):: __localStorageMixin(localStorage), - // Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified. - withFsType(fsType):: self + __localStorageMixin({ fsType: fsType }), - // The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...). - withPath(path):: self + __localStorageMixin({ path: path }), - }, - localType:: hidden.core.v1.localVolumeSource, - // A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options - withMountOptions(mountOptions):: self + if std.type(mountOptions) == 'array' then __specMixin({ mountOptions: mountOptions }) else __specMixin({ mountOptions: [mountOptions] }), - // A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options - withMountOptionsMixin(mountOptions):: self + if std.type(mountOptions) == 'array' then __specMixin({ mountOptions+: mountOptions }) else __specMixin({ mountOptions+: [mountOptions] }), - // NFS represents an NFS mount on the host. Provisioned by an admin. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - nfs:: { - local __nfsMixin(nfs) = __specMixin({ nfs+: nfs }), - mixinInstance(nfs):: __nfsMixin(nfs), - // Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withPath(path):: self + __nfsMixin({ path: path }), - // ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withReadOnly(readOnly):: self + __nfsMixin({ readOnly: readOnly }), - // Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withServer(server):: self + __nfsMixin({ server: server }), - }, - nfsType:: hidden.core.v1.nfsVolumeSource, - // NodeAffinity defines constraints that limit what nodes this volume can be accessed from. This field influences the scheduling of pods that use this volume. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __specMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // Required specifies hard node constraints that must be met. - required:: { - local __requiredMixin(required) = __nodeAffinityMixin({ required+: required }), - mixinInstance(required):: __requiredMixin(required), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.volumeNodeAffinity, - // What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming - withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy):: self + __specMixin({ persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy }), - // PhotonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine - photonPersistentDisk:: { - local __photonPersistentDiskMixin(photonPersistentDisk) = __specMixin({ photonPersistentDisk+: photonPersistentDisk }), - mixinInstance(photonPersistentDisk):: __photonPersistentDiskMixin(photonPersistentDisk), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __photonPersistentDiskMixin({ fsType: fsType }), - // ID that identifies Photon Controller persistent disk - withPdId(pdId):: self + __photonPersistentDiskMixin({ pdID: pdId }), - }, - photonPersistentDiskType:: hidden.core.v1.photonPersistentDiskVolumeSource, - // PortworxVolume represents a portworx volume attached and mounted on kubelets host machine - portworxVolume:: { - local __portworxVolumeMixin(portworxVolume) = __specMixin({ portworxVolume+: portworxVolume }), - mixinInstance(portworxVolume):: __portworxVolumeMixin(portworxVolume), - // FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __portworxVolumeMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __portworxVolumeMixin({ readOnly: readOnly }), - // VolumeID uniquely identifies a Portworx volume - withVolumeId(volumeId):: self + __portworxVolumeMixin({ volumeID: volumeId }), - }, - portworxVolumeType:: hidden.core.v1.portworxVolumeSource, - // Quobyte represents a Quobyte mount on the host that shares a pod's lifetime - quobyte:: { - local __quobyteMixin(quobyte) = __specMixin({ quobyte+: quobyte }), - mixinInstance(quobyte):: __quobyteMixin(quobyte), - // Group to map volume access to Default is no group - withGroup(group):: self + __quobyteMixin({ group: group }), - // ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false. - withReadOnly(readOnly):: self + __quobyteMixin({ readOnly: readOnly }), - // Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes - withRegistry(registry):: self + __quobyteMixin({ registry: registry }), - // Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin - withTenant(tenant):: self + __quobyteMixin({ tenant: tenant }), - // User to map volume access to Defaults to serivceaccount user - withUser(user):: self + __quobyteMixin({ user: user }), - // Volume is a string that references an already created Quobyte volume by name. - withVolume(volume):: self + __quobyteMixin({ volume: volume }), - }, - quobyteType:: hidden.core.v1.quobyteVolumeSource, - // RBD represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md - rbd:: { - local __rbdMixin(rbd) = __specMixin({ rbd+: rbd }), - mixinInstance(rbd):: __rbdMixin(rbd), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd - withFsType(fsType):: self + __rbdMixin({ fsType: fsType }), - // The rados image name. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withImage(image):: self + __rbdMixin({ image: image }), - // Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withKeyring(keyring):: self + __rbdMixin({ keyring: keyring }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors: monitors }) else __rbdMixin({ monitors: [monitors] }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors+: monitors }) else __rbdMixin({ monitors+: [monitors] }), - // The rados pool name. Default is rbd. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withPool(pool):: self + __rbdMixin({ pool: pool }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withReadOnly(readOnly):: self + __rbdMixin({ readOnly: readOnly }), - // SecretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __rbdMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // The rados user name. Default is admin. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withUser(user):: self + __rbdMixin({ user: user }), - }, - rbdType:: hidden.core.v1.rbdPersistentVolumeSource, - // ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes. - scaleIo:: { - local __scaleIoMixin(scaleIo) = __specMixin({ scaleIo+: scaleIo }), - mixinInstance(scaleIo):: __scaleIoMixin(scaleIo), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs" - withFsType(fsType):: self + __scaleIoMixin({ fsType: fsType }), - // The host address of the ScaleIO API Gateway. - withGateway(gateway):: self + __scaleIoMixin({ gateway: gateway }), - // The name of the ScaleIO Protection Domain for the configured storage. - withProtectionDomain(protectionDomain):: self + __scaleIoMixin({ protectionDomain: protectionDomain }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __scaleIoMixin({ readOnly: readOnly }), - // SecretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail. - secretRef:: { - local __secretRefMixin(secretRef) = __scaleIoMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // Flag to enable/disable SSL communication with Gateway, default false - withSslEnabled(sslEnabled):: self + __scaleIoMixin({ sslEnabled: sslEnabled }), - // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned. - withStorageMode(storageMode):: self + __scaleIoMixin({ storageMode: storageMode }), - // The ScaleIO Storage Pool associated with the protection domain. - withStoragePool(storagePool):: self + __scaleIoMixin({ storagePool: storagePool }), - // The name of the storage system as configured in ScaleIO. - withSystem(system):: self + __scaleIoMixin({ system: system }), - // The name of a volume already created in the ScaleIO system that is associated with this volume source. - withVolumeName(volumeName):: self + __scaleIoMixin({ volumeName: volumeName }), - }, - scaleIOType:: hidden.core.v1.scaleIoPersistentVolumeSource, - // Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass. - withStorageClassName(storageClassName):: self + __specMixin({ storageClassName: storageClassName }), - // StorageOS represents a StorageOS volume that is attached to the kubelet's host machine and mounted into the pod More info: https://releases.k8s.io/HEAD/examples/volumes/storageos/README.md - storageos:: { - local __storageosMixin(storageos) = __specMixin({ storageos+: storageos }), - mixinInstance(storageos):: __storageosMixin(storageos), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __storageosMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __storageosMixin({ readOnly: readOnly }), - // SecretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted. - secretRef:: { - local __secretRefMixin(secretRef) = __storageosMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __secretRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __secretRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __secretRefMixin({ uid: uid }), - }, - secretRefType:: hidden.core.v1.objectReference, - // VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace. - withVolumeName(volumeName):: self + __storageosMixin({ volumeName: volumeName }), - // VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to "default" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created. - withVolumeNamespace(volumeNamespace):: self + __storageosMixin({ volumeNamespace: volumeNamespace }), - }, - storageosType:: hidden.core.v1.storageOsPersistentVolumeSource, - // volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec. This is a beta feature. - withVolumeMode(volumeMode):: self + __specMixin({ volumeMode: volumeMode }), - // VsphereVolume represents a vSphere volume attached and mounted on kubelets host machine - vsphereVolume:: { - local __vsphereVolumeMixin(vsphereVolume) = __specMixin({ vsphereVolume+: vsphereVolume }), - mixinInstance(vsphereVolume):: __vsphereVolumeMixin(vsphereVolume), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __vsphereVolumeMixin({ fsType: fsType }), - // Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName. - withStoragePolicyId(storagePolicyId):: self + __vsphereVolumeMixin({ storagePolicyID: storagePolicyId }), - // Storage Policy Based Management (SPBM) profile name. - withStoragePolicyName(storagePolicyName):: self + __vsphereVolumeMixin({ storagePolicyName: storagePolicyName }), - // Path that identifies vSphere volume vmdk - withVolumePath(volumePath):: self + __vsphereVolumeMixin({ volumePath: volumePath }), - }, - vsphereVolumeType:: hidden.core.v1.vsphereVirtualDiskVolumeSource, - }, - specType:: hidden.core.v1.persistentVolumeSpec, - }, - }, - // PersistentVolumeClaim is a user's request for and claim to a persistent volume - persistentVolumeClaim:: { - local kind = { kind: 'PersistentVolumeClaim' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the desired characteristics of a volume requested by a pod author. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModes(accessModes):: self + if std.type(accessModes) == 'array' then __specMixin({ accessModes: accessModes }) else __specMixin({ accessModes: [accessModes] }), - // AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModesMixin(accessModes):: self + if std.type(accessModes) == 'array' then __specMixin({ accessModes+: accessModes }) else __specMixin({ accessModes+: [accessModes] }), - // This field requires the VolumeSnapshotDataSource alpha feature gate to be enabled and currently VolumeSnapshot is the only supported data source. If the provisioner can support VolumeSnapshot data source, it will create a new volume and data will be restored to the volume at the same time. If the provisioner does not support VolumeSnapshot data source, volume will not be created and the failure will be reported as an event. In the future, we plan to support more data source types and the behavior of the provisioner may change. - dataSource:: { - local __dataSourceMixin(dataSource) = __specMixin({ dataSource+: dataSource }), - mixinInstance(dataSource):: __dataSourceMixin(dataSource), - // APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required. - withApiGroup(apiGroup):: self + __dataSourceMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __dataSourceMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __dataSourceMixin({ name: name }), - }, - dataSourceType:: hidden.core.v1.typedLocalObjectReference, - // Resources represents the minimum resources the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources - resources:: { - local __resourcesMixin(resources) = __specMixin({ resources+: resources }), - mixinInstance(resources):: __resourcesMixin(resources), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimits(limits):: self + __resourcesMixin({ limits: limits }), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimitsMixin(limits):: self + __resourcesMixin({ limits+: limits }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequests(requests):: self + __resourcesMixin({ requests: requests }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequestsMixin(requests):: self + __resourcesMixin({ requests+: requests }), - }, - resourcesType:: hidden.core.v1.resourceRequirements, - // A label query over volumes to consider for binding. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1 - withStorageClassName(storageClassName):: self + __specMixin({ storageClassName: storageClassName }), - // volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec. This is a beta feature. - withVolumeMode(volumeMode):: self + __specMixin({ volumeMode: volumeMode }), - // VolumeName is the binding reference to the PersistentVolume backing this claim. - withVolumeName(volumeName):: self + __specMixin({ volumeName: volumeName }), - }, - specType:: hidden.core.v1.persistentVolumeClaimSpec, - }, - }, - // Pod is a collection of containers that can run on a host. This resource is created by clients and scheduled onto hosts. - pod:: { - local kind = { kind: 'Pod' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - }, - // PodTemplate describes a template for creating copies of a predefined pod. - podTemplate:: { - local kind = { kind: 'PodTemplate' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Template defines the pods that will be created from this pod template. https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // ReplicationController represents the configuration of a replication controller. - replicationController:: { - local kind = { kind: 'ReplicationController' }, - new():: apiVersion + kind, - mixin:: { - // If the Labels of a ReplicationController are empty, they are defaulted to be the same as the Pod(s) that the replication controller manages. Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the specification of the desired behavior of the replication controller. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withSelector(selector):: self + __specMixin({ selector: selector }), - // Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withSelectorMixin(selector):: self + __specMixin({ selector+: selector }), - // Template is the object that describes the pod that will be created if insufficient replicas are detected. This takes precedence over a TemplateRef. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.core.v1.replicationControllerSpec, - }, - }, - // ResourceQuota sets aggregate quota restrictions enforced per namespace - resourceQuota:: { - local kind = { kind: 'ResourceQuota' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the desired quota. https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHard(hard):: self + __specMixin({ hard: hard }), - // hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHardMixin(hard):: self + __specMixin({ hard+: hard }), - // scopeSelector is also a collection of filters like scopes that must match each object tracked by a quota but expressed using ScopeSelectorOperator in combination with possible values. For a resource to match, both scopes AND scopeSelector (if specified in spec), must be matched. - scopeSelector:: { - local __scopeSelectorMixin(scopeSelector) = __specMixin({ scopeSelector+: scopeSelector }), - mixinInstance(scopeSelector):: __scopeSelectorMixin(scopeSelector), - // A list of scope selector requirements by scope of the resources. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __scopeSelectorMixin({ matchExpressions: matchExpressions }) else __scopeSelectorMixin({ matchExpressions: [matchExpressions] }), - // A list of scope selector requirements by scope of the resources. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __scopeSelectorMixin({ matchExpressions+: matchExpressions }) else __scopeSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.core.v1.scopedResourceSelectorRequirement, - }, - scopeSelectorType:: hidden.core.v1.scopeSelector, - // A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects. - withScopes(scopes):: self + if std.type(scopes) == 'array' then __specMixin({ scopes: scopes }) else __specMixin({ scopes: [scopes] }), - // A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects. - withScopesMixin(scopes):: self + if std.type(scopes) == 'array' then __specMixin({ scopes+: scopes }) else __specMixin({ scopes+: [scopes] }), - }, - specType:: hidden.core.v1.resourceQuotaSpec, - }, - }, - // Secret holds secret data of a certain type. The total bytes of the values in the Data field must be less than MaxSecretSize bytes. - secret:: { - local kind = { kind: 'Secret' }, - new(name='', data='', type='Opaque'):: apiVersion + kind + self.withData(data).withType(type) + self.mixin.metadata.withName(name), - // Data contains the secret data. Each key must consist of alphanumeric characters, '-', '_' or '.'. The serialized form of the secret data is a base64 encoded string, representing the arbitrary (possibly non-string) data value here. Described in https://tools.ietf.org/html/rfc4648#section-4 - withData(data):: self + { data: data }, - // Data contains the secret data. Each key must consist of alphanumeric characters, '-', '_' or '.'. The serialized form of the secret data is a base64 encoded string, representing the arbitrary (possibly non-string) data value here. Described in https://tools.ietf.org/html/rfc4648#section-4 - withDataMixin(data):: self + { data+: data }, - // stringData allows specifying non-binary secret data in string form. It is provided as a write-only convenience method. All keys and values are merged into the data field on write, overwriting any existing values. It is never output when reading from the API. - withStringData(stringData):: self + { stringData: stringData }, - // stringData allows specifying non-binary secret data in string form. It is provided as a write-only convenience method. All keys and values are merged into the data field on write, overwriting any existing values. It is never output when reading from the API. - withStringDataMixin(stringData):: self + { stringData+: stringData }, - // Used to facilitate programmatic handling of secret data. - withType(type):: self + { type: type }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // Service is a named abstraction of software service (for example, mysql) consisting of local port (for example 3306) that the proxy listens on, and the selector that determines which pods will answer requests sent through the proxy. - service:: { - local kind = { kind: 'Service' }, - new(name='', selector='', ports=''):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withPorts(ports).withSelector(selector), - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the behavior of a service. https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // clusterIP is the IP address of the service and is usually assigned randomly by the master. If an address is specified manually and is not in use by others, it will be allocated to the service; otherwise, creation of the service will fail. This field can not be changed through updates. Valid values are "None", empty string (""), or a valid IP address. "None" can be specified for headless services when proxying is not required. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withClusterIp(clusterIp):: self + __specMixin({ clusterIP: clusterIp }), - // externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. - withExternalIps(externalIps):: self + if std.type(externalIps) == 'array' then __specMixin({ externalIPs: externalIps }) else __specMixin({ externalIPs: [externalIps] }), - // externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. - withExternalIpsMixin(externalIps):: self + if std.type(externalIps) == 'array' then __specMixin({ externalIPs+: externalIps }) else __specMixin({ externalIPs+: [externalIps] }), - // externalName is the external reference that kubedns or equivalent will return as a CNAME record for this service. No proxying will be involved. Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123) and requires Type to be ExternalName. - withExternalName(externalName):: self + __specMixin({ externalName: externalName }), - // externalTrafficPolicy denotes if this Service desires to route external traffic to node-local or cluster-wide endpoints. "Local" preserves the client source IP and avoids a second hop for LoadBalancer and Nodeport type services, but risks potentially imbalanced traffic spreading. "Cluster" obscures the client source IP and may cause a second hop to another node, but should have good overall load-spreading. - withExternalTrafficPolicy(externalTrafficPolicy):: self + __specMixin({ externalTrafficPolicy: externalTrafficPolicy }), - // healthCheckNodePort specifies the healthcheck nodePort for the service. If not specified, HealthCheckNodePort is created by the service api backend with the allocated nodePort. Will use user-specified nodePort value if specified by the client. Only effects when Type is set to LoadBalancer and ExternalTrafficPolicy is set to Local. - withHealthCheckNodePort(healthCheckNodePort):: self + __specMixin({ healthCheckNodePort: healthCheckNodePort }), - // Only applies to Service Type: LoadBalancer LoadBalancer will get created with the IP specified in this field. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature. - withLoadBalancerIp(loadBalancerIp):: self + __specMixin({ loadBalancerIP: loadBalancerIp }), - // If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/ - withLoadBalancerSourceRanges(loadBalancerSourceRanges):: self + if std.type(loadBalancerSourceRanges) == 'array' then __specMixin({ loadBalancerSourceRanges: loadBalancerSourceRanges }) else __specMixin({ loadBalancerSourceRanges: [loadBalancerSourceRanges] }), - // If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/ - withLoadBalancerSourceRangesMixin(loadBalancerSourceRanges):: self + if std.type(loadBalancerSourceRanges) == 'array' then __specMixin({ loadBalancerSourceRanges+: loadBalancerSourceRanges }) else __specMixin({ loadBalancerSourceRanges+: [loadBalancerSourceRanges] }), - // The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withPorts(ports):: self + if std.type(ports) == 'array' then __specMixin({ ports: ports }) else __specMixin({ ports: [ports] }), - // The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withPortsMixin(ports):: self + if std.type(ports) == 'array' then __specMixin({ ports+: ports }) else __specMixin({ ports+: [ports] }), - portsType:: hidden.core.v1.servicePort, - // publishNotReadyAddresses, when set to true, indicates that DNS implementations must publish the notReadyAddresses of subsets for the Endpoints associated with the Service. The default value is false. The primary use case for setting this field is to use a StatefulSet's Headless Service to propagate SRV records for its Pods without respect to their readiness for purpose of peer discovery. - withPublishNotReadyAddresses(publishNotReadyAddresses):: self + __specMixin({ publishNotReadyAddresses: publishNotReadyAddresses }), - // Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/ - withSelector(selector):: self + __specMixin({ selector: selector }), - // Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/ - withSelectorMixin(selector):: self + __specMixin({ selector+: selector }), - // Supports "ClientIP" and "None". Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withSessionAffinity(sessionAffinity):: self + __specMixin({ sessionAffinity: sessionAffinity }), - // sessionAffinityConfig contains the configurations of session affinity. - sessionAffinityConfig:: { - local __sessionAffinityConfigMixin(sessionAffinityConfig) = __specMixin({ sessionAffinityConfig+: sessionAffinityConfig }), - mixinInstance(sessionAffinityConfig):: __sessionAffinityConfigMixin(sessionAffinityConfig), - // clientIP contains the configurations of Client IP based session affinity. - clientIp:: { - local __clientIpMixin(clientIp) = __sessionAffinityConfigMixin({ clientIp+: clientIp }), - mixinInstance(clientIp):: __clientIpMixin(clientIp), - // timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours). - withTimeoutSeconds(timeoutSeconds):: self + __clientIpMixin({ timeoutSeconds: timeoutSeconds }), - }, - clientIPType:: hidden.core.v1.clientIpConfig, - }, - sessionAffinityConfigType:: hidden.core.v1.sessionAffinityConfig, - // type determines how the Service is exposed. Defaults to ClusterIP. Valid options are ExternalName, ClusterIP, NodePort, and LoadBalancer. "ExternalName" maps to the specified externalName. "ClusterIP" allocates a cluster-internal IP address for load-balancing to endpoints. Endpoints are determined by the selector or if that is not specified, by manual construction of an Endpoints object. If clusterIP is "None", no virtual IP is allocated and the endpoints are published as a set of endpoints rather than a stable IP. "NodePort" builds on ClusterIP and allocates a port on every node which routes to the clusterIP. "LoadBalancer" builds on NodePort and creates an external load-balancer (if supported in the current cloud) which routes to the clusterIP. More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types - withType(type):: self + __specMixin({ type: type }), - }, - specType:: hidden.core.v1.serviceSpec, - }, - }, - // ServiceAccount binds together: * a name, understood by users, and perhaps by peripheral systems, for an identity * a principal that can be authenticated and authorized * a set of secrets - serviceAccount:: { - local kind = { kind: 'ServiceAccount' }, - new(name=''):: apiVersion + kind + self.mixin.metadata.withName(name), - // AutomountServiceAccountToken indicates whether pods running as this service account should have an API token automatically mounted. Can be overridden at the pod level. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + { automountServiceAccountToken: automountServiceAccountToken }, - // ImagePullSecrets is a list of references to secrets in the same namespace to use for pulling any images in pods that reference this ServiceAccount. ImagePullSecrets are distinct from Secrets because Secrets can be mounted in the pod, but ImagePullSecrets are only accessed by the kubelet. More info: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then { imagePullSecrets: imagePullSecrets } else { imagePullSecrets: [imagePullSecrets] }, - // ImagePullSecrets is a list of references to secrets in the same namespace to use for pulling any images in pods that reference this ServiceAccount. ImagePullSecrets are distinct from Secrets because Secrets can be mounted in the pod, but ImagePullSecrets are only accessed by the kubelet. More info: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then { imagePullSecrets+: imagePullSecrets } else { imagePullSecrets+: [imagePullSecrets] }, - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // Secrets is the list of secrets allowed to be used by pods running using this ServiceAccount. More info: https://kubernetes.io/docs/concepts/configuration/secret - withSecrets(secrets):: self + if std.type(secrets) == 'array' then { secrets: secrets } else { secrets: [secrets] }, - // Secrets is the list of secrets allowed to be used by pods running using this ServiceAccount. More info: https://kubernetes.io/docs/concepts/configuration/secret - withSecretsMixin(secrets):: self + if std.type(secrets) == 'array' then { secrets+: secrets } else { secrets+: [secrets] }, - secretsType:: hidden.core.v1.objectReference, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - }, - }, - events:: { - v1beta1:: { - local apiVersion = { apiVersion: 'events.k8s.io/v1beta1' }, - // Event is a report of an event somewhere in the cluster. It generally denotes some state change in the system. - event:: { - local kind = { kind: 'Event' }, - new():: apiVersion + kind, - // What action was taken/failed regarding to the regarding object. - withAction(action):: self + { action: action }, - // Deprecated field assuring backward compatibility with core.v1 Event type - withDeprecatedCount(deprecatedCount):: self + { deprecatedCount: deprecatedCount }, - // Deprecated field assuring backward compatibility with core.v1 Event type - withDeprecatedFirstTimestamp(deprecatedFirstTimestamp):: self + { deprecatedFirstTimestamp: deprecatedFirstTimestamp }, - // Deprecated field assuring backward compatibility with core.v1 Event type - withDeprecatedLastTimestamp(deprecatedLastTimestamp):: self + { deprecatedLastTimestamp: deprecatedLastTimestamp }, - // Required. Time when this Event was first observed. - withEventTime(eventTime):: self + { eventTime: eventTime }, - // Optional. A human-readable description of the status of this operation. Maximal length of the note is 1kB, but libraries should be prepared to handle values up to 64kB. - withNote(note):: self + { note: note }, - // Why the action was taken. - withReason(reason):: self + { reason: reason }, - // Name of the controller that emitted this Event, e.g. `kubernetes.io/kubelet`. - withReportingController(reportingController):: self + { reportingController: reportingController }, - // ID of the controller instance, e.g. `kubelet-xyzf`. - withReportingInstance(reportingInstance):: self + { reportingInstance: reportingInstance }, - // Type of this event (Normal, Warning), new types could be added in the future. - withType(type):: self + { type: type }, - mixin:: { - // Deprecated field assuring backward compatibility with core.v1 Event type - deprecatedSource:: { - local __deprecatedSourceMixin(deprecatedSource) = { deprecatedSource+: deprecatedSource }, - mixinInstance(deprecatedSource):: __deprecatedSourceMixin(deprecatedSource), - // Component from which the event is generated. - withComponent(component):: self + __deprecatedSourceMixin({ component: component }), - // Node name on which the event is generated. - withHost(host):: self + __deprecatedSourceMixin({ host: host }), - }, - deprecatedSourceType:: hidden.core.v1.eventSource, - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The object this Event is about. In most cases it's an Object reporting controller implements. E.g. ReplicaSetController implements ReplicaSets and this event is emitted because it acts on some changes in a ReplicaSet object. - regarding:: { - local __regardingMixin(regarding) = { regarding+: regarding }, - mixinInstance(regarding):: __regardingMixin(regarding), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __regardingMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __regardingMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __regardingMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __regardingMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __regardingMixin({ uid: uid }), - }, - regardingType:: hidden.core.v1.objectReference, - // Optional secondary object for more complex actions. E.g. when regarding object triggers a creation or deletion of related object. - related:: { - local __relatedMixin(related) = { related+: related }, - mixinInstance(related):: __relatedMixin(related), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __relatedMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __relatedMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __relatedMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __relatedMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __relatedMixin({ uid: uid }), - }, - relatedType:: hidden.core.v1.objectReference, - // Data about the Event series this event represents or nil if it's a singleton Event. - series:: { - local __seriesMixin(series) = { series+: series }, - mixinInstance(series):: __seriesMixin(series), - // Number of occurrences in this series up to the last heartbeat time - withCount(count):: self + __seriesMixin({ count: count }), - // Time when last Event from the series was seen before last heartbeat. - withLastObservedTime(lastObservedTime):: self + __seriesMixin({ lastObservedTime: lastObservedTime }), - // Information whether this series is ongoing or finished. - withState(state):: self + __seriesMixin({ state: state }), - }, - seriesType:: hidden.events.v1beta1.eventSeries, - }, - }, - }, - }, - extensions:: { - v1beta1:: { - local apiVersion = { apiVersion: 'extensions/v1beta1' }, - // DEPRECATED - This group version of DaemonSet is deprecated by apps/v1beta2/DaemonSet. See the release notes for more information. DaemonSet represents the configuration of a daemon set. - daemonSet:: { - local kind = { kind: 'DaemonSet' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // The desired behavior of this daemon set. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. If empty, defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // DEPRECATED. A sequence number representing a specific generation of the template. Populated by the system. It can be set only during the creation. - withTemplateGeneration(templateGeneration):: self + __specMixin({ templateGeneration: templateGeneration }), - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = __specMixin({ updateStrategy+: updateStrategy }), - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is OnDelete. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.extensions.v1beta1.daemonSetUpdateStrategy, - }, - specType:: hidden.extensions.v1beta1.daemonSetSpec, - }, - }, - // DEPRECATED - This group version of Deployment is deprecated by apps/v1beta2/Deployment. See the release notes for more information. Deployment enables declarative updates for Pods and ReplicaSets. - deployment:: { - local kind = { kind: 'Deployment' }, - new(name='', replicas=1, containers='', podLabels={ app: 'name' }):: apiVersion + kind + self.mixin.metadata.withName(name) + self.mixin.spec.withReplicas(replicas) + self.mixin.spec.template.metadata.withLabels(podLabels) + self.mixin.spec.template.spec.withContainers(containers), - mixin:: { - // Standard object metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the Deployment. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Indicates that the deployment is paused and will not be processed by the deployment controller. - withPaused(paused):: self + __specMixin({ paused: paused }), - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. This is set to the max value of int32 (i.e. 2147483647) by default, which means "no deadline". - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + __specMixin({ progressDeadlineSeconds: progressDeadlineSeconds }), - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. This is set to the max value of int32 (i.e. 2147483647) by default, which means "retaining all old RelicaSets". - withRevisionHistoryLimit(revisionHistoryLimit):: self + __specMixin({ revisionHistoryLimit: revisionHistoryLimit }), - // DEPRECATED. The config this deployment is rolling back to. Will be cleared after rollback is done. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = __specMixin({ rollbackTo+: rollbackTo }), - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.extensions.v1beta1.rollbackConfig, - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = __specMixin({ strategy+: strategy }), - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. By default, a value of 1 is used. Example: when this is set to 30%, the new RC can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new RC can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. By default, a fixed value of 1 is used. Example: when this is set to 30%, the old RC can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old RC can be scaled down further, followed by scaling up the new RC, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.extensions.v1beta1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.extensions.v1beta1.deploymentSpec, - }, - }, - // DEPRECATED. DeploymentRollback stores the information required to rollback a deployment. - deploymentRollback:: { - local kind = { kind: 'DeploymentRollback' }, - new(name=''):: apiVersion + kind + self.withName(name), - // Required: This must match the Name of a deployment. - withName(name):: self + { name: name }, - // The annotations to be updated to a deployment - withUpdatedAnnotations(updatedAnnotations):: self + { updatedAnnotations: updatedAnnotations }, - // The annotations to be updated to a deployment - withUpdatedAnnotationsMixin(updatedAnnotations):: self + { updatedAnnotations+: updatedAnnotations }, - mixin:: { - // The config of this deployment rollback. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = { rollbackTo+: rollbackTo }, - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.extensions.v1beta1.rollbackConfig, - }, - }, - // Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc. DEPRECATED - This group version of Ingress is deprecated by networking.k8s.io/v1beta1 Ingress. See the release notes for more information. - ingress:: { - local kind = { kind: 'Ingress' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec is the desired state of the Ingress. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default. - backend:: { - local __backendMixin(backend) = __specMixin({ backend+: backend }), - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.extensions.v1beta1.ingressBackend, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRules(rules):: self + if std.type(rules) == 'array' then __specMixin({ rules: rules }) else __specMixin({ rules: [rules] }), - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRulesMixin(rules):: self + if std.type(rules) == 'array' then __specMixin({ rules+: rules }) else __specMixin({ rules+: [rules] }), - rulesType:: hidden.extensions.v1beta1.ingressRule, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTls(tls):: self + if std.type(tls) == 'array' then __specMixin({ tls: tls }) else __specMixin({ tls: [tls] }), - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTlsMixin(tls):: self + if std.type(tls) == 'array' then __specMixin({ tls+: tls }) else __specMixin({ tls+: [tls] }), - tlsType:: hidden.extensions.v1beta1.ingressTls, - }, - specType:: hidden.extensions.v1beta1.ingressSpec, - }, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicy is deprecated by networking/v1/NetworkPolicy. NetworkPolicy describes what network traffic is allowed for a set of Pods - networkPolicy:: { - local kind = { kind: 'NetworkPolicy' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior for this NetworkPolicy. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgress(egress):: self + if std.type(egress) == 'array' then __specMixin({ egress: egress }) else __specMixin({ egress: [egress] }), - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgressMixin(egress):: self + if std.type(egress) == 'array' then __specMixin({ egress+: egress }) else __specMixin({ egress+: [egress] }), - egressType:: hidden.extensions.v1beta1.networkPolicyEgressRule, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default). - withIngress(ingress):: self + if std.type(ingress) == 'array' then __specMixin({ ingress: ingress }) else __specMixin({ ingress: [ingress] }), - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default). - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then __specMixin({ ingress+: ingress }) else __specMixin({ ingress+: [ingress] }), - ingressType:: hidden.extensions.v1beta1.networkPolicyIngressRule, - // Selects the pods to which this NetworkPolicy object applies. The array of ingress rules is applied to any pods selected by this field. Multiple network policies can select the same set of pods. In this case, the ingress rules for each are combined additively. This field is NOT optional and follows standard label selector semantics. An empty podSelector matches all pods in this namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = __specMixin({ podSelector+: podSelector }), - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypes(policyTypes):: self + if std.type(policyTypes) == 'array' then __specMixin({ policyTypes: policyTypes }) else __specMixin({ policyTypes: [policyTypes] }), - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypesMixin(policyTypes):: self + if std.type(policyTypes) == 'array' then __specMixin({ policyTypes+: policyTypes }) else __specMixin({ policyTypes+: [policyTypes] }), - }, - specType:: hidden.extensions.v1beta1.networkPolicySpec, - }, - }, - // PodSecurityPolicy governs the ability to make requests that affect the Security Context that will be applied to a pod and container. Deprecated: use PodSecurityPolicy from policy API Group instead. - podSecurityPolicy:: { - local kind = { kind: 'PodSecurityPolicy' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // spec defines the policy enforced. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true. - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + __specMixin({ allowPrivilegeEscalation: allowPrivilegeEscalation }), - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDrivers(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then __specMixin({ allowedCSIDrivers: allowedCsiDrivers }) else __specMixin({ allowedCSIDrivers: [allowedCsiDrivers] }), - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDriversMixin(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then __specMixin({ allowedCSIDrivers+: allowedCsiDrivers }) else __specMixin({ allowedCSIDrivers+: [allowedCsiDrivers] }), - allowedCSIDriversType:: hidden.extensions.v1beta1.allowedCsiDriver, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilities(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then __specMixin({ allowedCapabilities: allowedCapabilities }) else __specMixin({ allowedCapabilities: [allowedCapabilities] }), - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilitiesMixin(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then __specMixin({ allowedCapabilities+: allowedCapabilities }) else __specMixin({ allowedCapabilities+: [allowedCapabilities] }), - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumes(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then __specMixin({ allowedFlexVolumes: allowedFlexVolumes }) else __specMixin({ allowedFlexVolumes: [allowedFlexVolumes] }), - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumesMixin(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then __specMixin({ allowedFlexVolumes+: allowedFlexVolumes }) else __specMixin({ allowedFlexVolumes+: [allowedFlexVolumes] }), - allowedFlexVolumesType:: hidden.extensions.v1beta1.allowedFlexVolume, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPaths(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then __specMixin({ allowedHostPaths: allowedHostPaths }) else __specMixin({ allowedHostPaths: [allowedHostPaths] }), - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPathsMixin(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then __specMixin({ allowedHostPaths+: allowedHostPaths }) else __specMixin({ allowedHostPaths+: [allowedHostPaths] }), - allowedHostPathsType:: hidden.extensions.v1beta1.allowedHostPath, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypes(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then __specMixin({ allowedProcMountTypes: allowedProcMountTypes }) else __specMixin({ allowedProcMountTypes: [allowedProcMountTypes] }), - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypesMixin(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then __specMixin({ allowedProcMountTypes+: allowedProcMountTypes }) else __specMixin({ allowedProcMountTypes+: [allowedProcMountTypes] }), - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctls(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then __specMixin({ allowedUnsafeSysctls: allowedUnsafeSysctls }) else __specMixin({ allowedUnsafeSysctls: [allowedUnsafeSysctls] }), - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then __specMixin({ allowedUnsafeSysctls+: allowedUnsafeSysctls }) else __specMixin({ allowedUnsafeSysctls+: [allowedUnsafeSysctls] }), - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilities(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then __specMixin({ defaultAddCapabilities: defaultAddCapabilities }) else __specMixin({ defaultAddCapabilities: [defaultAddCapabilities] }), - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilitiesMixin(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then __specMixin({ defaultAddCapabilities+: defaultAddCapabilities }) else __specMixin({ defaultAddCapabilities+: [defaultAddCapabilities] }), - // defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process. - withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation):: self + __specMixin({ defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation }), - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctls(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then __specMixin({ forbiddenSysctls: forbiddenSysctls }) else __specMixin({ forbiddenSysctls: [forbiddenSysctls] }), - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctlsMixin(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then __specMixin({ forbiddenSysctls+: forbiddenSysctls }) else __specMixin({ forbiddenSysctls+: [forbiddenSysctls] }), - // fsGroup is the strategy that will dictate what fs group is used by the SecurityContext. - fsGroup:: { - local __fsGroupMixin(fsGroup) = __specMixin({ fsGroup+: fsGroup }), - mixinInstance(fsGroup):: __fsGroupMixin(fsGroup), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges: ranges }) else __fsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges+: ranges }) else __fsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + __fsGroupMixin({ rule: rule }), - }, - fsGroupType:: hidden.extensions.v1beta1.fsGroupStrategyOptions, - // hostIPC determines if the policy allows the use of HostIPC in the pod spec. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // hostNetwork determines if the policy allows the use of HostNetwork in the pod spec. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // hostPID determines if the policy allows the use of HostPID in the pod spec. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPorts(hostPorts):: self + if std.type(hostPorts) == 'array' then __specMixin({ hostPorts: hostPorts }) else __specMixin({ hostPorts: [hostPorts] }), - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPortsMixin(hostPorts):: self + if std.type(hostPorts) == 'array' then __specMixin({ hostPorts+: hostPorts }) else __specMixin({ hostPorts+: [hostPorts] }), - hostPortsType:: hidden.extensions.v1beta1.hostPortRange, - // privileged determines if a pod can request to be run as privileged. - withPrivileged(privileged):: self + __specMixin({ privileged: privileged }), - // readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to. - withReadOnlyRootFilesystem(readOnlyRootFilesystem):: self + __specMixin({ readOnlyRootFilesystem: readOnlyRootFilesystem }), - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilities(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then __specMixin({ requiredDropCapabilities: requiredDropCapabilities }) else __specMixin({ requiredDropCapabilities: [requiredDropCapabilities] }), - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilitiesMixin(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then __specMixin({ requiredDropCapabilities+: requiredDropCapabilities }) else __specMixin({ requiredDropCapabilities+: [requiredDropCapabilities] }), - // RunAsGroup is the strategy that will dictate the allowable RunAsGroup values that may be set. If this field is omitted, the pod's RunAsGroup can take any value. This field requires the RunAsGroup feature gate to be enabled. - runAsGroup:: { - local __runAsGroupMixin(runAsGroup) = __specMixin({ runAsGroup+: runAsGroup }), - mixinInstance(runAsGroup):: __runAsGroupMixin(runAsGroup), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges: ranges }) else __runAsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges+: ranges }) else __runAsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + __runAsGroupMixin({ rule: rule }), - }, - runAsGroupType:: hidden.extensions.v1beta1.runAsGroupStrategyOptions, - // runAsUser is the strategy that will dictate the allowable RunAsUser values that may be set. - runAsUser:: { - local __runAsUserMixin(runAsUser) = __specMixin({ runAsUser+: runAsUser }), - mixinInstance(runAsUser):: __runAsUserMixin(runAsUser), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges: ranges }) else __runAsUserMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges+: ranges }) else __runAsUserMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + __runAsUserMixin({ rule: rule }), - }, - runAsUserType:: hidden.extensions.v1beta1.runAsUserStrategyOptions, - // seLinux is the strategy that will dictate the allowable labels that may be set. - seLinux:: { - local __seLinuxMixin(seLinux) = __specMixin({ seLinux+: seLinux }), - mixinInstance(seLinux):: __seLinuxMixin(seLinux), - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + __seLinuxMixin({ rule: rule }), - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __seLinuxMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - seLinuxType:: hidden.extensions.v1beta1.seLinuxStrategyOptions, - // supplementalGroups is the strategy that will dictate what supplemental groups are used by the SecurityContext. - supplementalGroups:: { - local __supplementalGroupsMixin(supplementalGroups) = __specMixin({ supplementalGroups+: supplementalGroups }), - mixinInstance(supplementalGroups):: __supplementalGroupsMixin(supplementalGroups), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges: ranges }) else __supplementalGroupsMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges+: ranges }) else __supplementalGroupsMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + __supplementalGroupsMixin({ rule: rule }), - }, - supplementalGroupsType:: hidden.extensions.v1beta1.supplementalGroupsStrategyOptions, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - }, - specType:: hidden.extensions.v1beta1.podSecurityPolicySpec, - }, - }, - // DEPRECATED - This group version of ReplicaSet is deprecated by apps/v1beta2/ReplicaSet. See the release notes for more information. ReplicaSet ensures that a specified number of pod replicas are running at any given time. - replicaSet:: { - local kind = { kind: 'ReplicaSet' }, - new():: apiVersion + kind, - mixin:: { - // If the Labels of a ReplicaSet are empty, they are defaulted to be the same as the Pod(s) that the ReplicaSet manages. Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec defines the specification of the desired behavior of the ReplicaSet. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + __specMixin({ minReadySeconds: minReadySeconds }), - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - // Selector is a label query over pods that should match the replica count. If the selector is empty, it is defaulted to the labels present on the pod template. Label keys and values that must match in order to be controlled by this replica set. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - specType:: hidden.extensions.v1beta1.replicaSetSpec, - }, - }, - // represents a scaling request for a resource. - scale:: { - local kind = { kind: 'Scale' }, - new(replicas=1):: apiVersion + kind + self.mixin.spec.withReplicas(replicas), - mixin:: { - // Standard object metadata; More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // defines the behavior of the scale. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // desired number of instances for the scaled object. - withReplicas(replicas):: self + __specMixin({ replicas: replicas }), - }, - specType:: hidden.extensions.v1beta1.scaleSpec, - }, - }, - }, - }, - meta:: { - v1:: { - local apiVersion = { apiVersion: 'storage.k8s.io/v1' }, - // Patch is provided to give a concrete name and type to the Kubernetes PATCH request body. - volumeAttachment:: { - local kind = { kind: 'VolumeAttachment' }, - new():: apiVersion + kind, - mixin:: {}, - }, - }, - }, - networking:: { - v1:: { - local apiVersion = { apiVersion: 'networking.k8s.io/v1' }, - // NetworkPolicy describes what network traffic is allowed for a set of Pods - networkPolicy:: { - local kind = { kind: 'NetworkPolicy' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior for this NetworkPolicy. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgress(egress):: self + if std.type(egress) == 'array' then __specMixin({ egress: egress }) else __specMixin({ egress: [egress] }), - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgressMixin(egress):: self + if std.type(egress) == 'array' then __specMixin({ egress+: egress }) else __specMixin({ egress+: [egress] }), - egressType:: hidden.networking.v1.networkPolicyEgressRule, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default) - withIngress(ingress):: self + if std.type(ingress) == 'array' then __specMixin({ ingress: ingress }) else __specMixin({ ingress: [ingress] }), - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default) - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then __specMixin({ ingress+: ingress }) else __specMixin({ ingress+: [ingress] }), - ingressType:: hidden.networking.v1.networkPolicyIngressRule, - // Selects the pods to which this NetworkPolicy object applies. The array of ingress rules is applied to any pods selected by this field. Multiple network policies can select the same set of pods. In this case, the ingress rules for each are combined additively. This field is NOT optional and follows standard label selector semantics. An empty podSelector matches all pods in this namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = __specMixin({ podSelector+: podSelector }), - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypes(policyTypes):: self + if std.type(policyTypes) == 'array' then __specMixin({ policyTypes: policyTypes }) else __specMixin({ policyTypes: [policyTypes] }), - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypesMixin(policyTypes):: self + if std.type(policyTypes) == 'array' then __specMixin({ policyTypes+: policyTypes }) else __specMixin({ policyTypes+: [policyTypes] }), - }, - specType:: hidden.networking.v1.networkPolicySpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'networking.k8s.io/v1beta1' }, - // Ingress is a collection of rules that allow inbound connections to reach the endpoints defined by a backend. An Ingress can be configured to give services externally-reachable urls, load balance traffic, terminate SSL, offer name based virtual hosting etc. - ingress:: { - local kind = { kind: 'Ingress' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Spec is the desired state of the Ingress. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default. - backend:: { - local __backendMixin(backend) = __specMixin({ backend+: backend }), - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.networking.v1beta1.ingressBackend, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRules(rules):: self + if std.type(rules) == 'array' then __specMixin({ rules: rules }) else __specMixin({ rules: [rules] }), - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRulesMixin(rules):: self + if std.type(rules) == 'array' then __specMixin({ rules+: rules }) else __specMixin({ rules+: [rules] }), - rulesType:: hidden.networking.v1beta1.ingressRule, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTls(tls):: self + if std.type(tls) == 'array' then __specMixin({ tls: tls }) else __specMixin({ tls: [tls] }), - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTlsMixin(tls):: self + if std.type(tls) == 'array' then __specMixin({ tls+: tls }) else __specMixin({ tls+: [tls] }), - tlsType:: hidden.networking.v1beta1.ingressTls, - }, - specType:: hidden.networking.v1beta1.ingressSpec, - }, - }, - }, - }, - node:: { - v1beta1:: { - local apiVersion = { apiVersion: 'node.k8s.io/v1beta1' }, - // RuntimeClass defines a class of container runtime supported in the cluster. The RuntimeClass is used to determine which container runtime is used to run all containers in a pod. RuntimeClasses are (currently) manually defined by a user or cluster provisioner, and referenced in the PodSpec. The Kubelet is responsible for resolving the RuntimeClassName reference before running the pod. For more details, see https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md - runtimeClass:: { - local kind = { kind: 'RuntimeClass' }, - new():: apiVersion + kind, - // Handler specifies the underlying runtime and configuration that the CRI implementation will use to handle pods of this class. The possible values are specific to the node & CRI configuration. It is assumed that all handlers are available on every node, and handlers of the same name are equivalent on every node. For example, a handler called "runc" might specify that the runc OCI runtime (using native Linux containers) will be used to run the containers in a pod. The Handler must conform to the DNS Label (RFC 1123) requirements, and is immutable. - withHandler(handler):: self + { handler: handler }, - mixin:: { - // More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - }, - }, - policy:: { - v1beta1:: { - local apiVersion = { apiVersion: 'policy/v1beta1' }, - // Eviction evicts a pod from its node subject to certain policies and safety constraints. This is a subresource of Pod. A request to cause such an eviction is created by POSTing to .../pods//evictions. - eviction:: { - local kind = { kind: 'Eviction' }, - new():: apiVersion + kind, - mixin:: { - // DeleteOptions may be provided - deleteOptions:: { - local __deleteOptionsMixin(deleteOptions) = { deleteOptions+: deleteOptions }, - mixinInstance(deleteOptions):: __deleteOptionsMixin(deleteOptions), - // When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed - withDryRun(dryRun):: self + if std.type(dryRun) == 'array' then __deleteOptionsMixin({ dryRun: dryRun }) else __deleteOptionsMixin({ dryRun: [dryRun] }), - // When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed - withDryRunMixin(dryRun):: self + if std.type(dryRun) == 'array' then __deleteOptionsMixin({ dryRun+: dryRun }) else __deleteOptionsMixin({ dryRun+: [dryRun] }), - // The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. - withGracePeriodSeconds(gracePeriodSeconds):: self + __deleteOptionsMixin({ gracePeriodSeconds: gracePeriodSeconds }), - // Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the "orphan" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. - withOrphanDependents(orphanDependents):: self + __deleteOptionsMixin({ orphanDependents: orphanDependents }), - // Must be fulfilled before a deletion is carried out. If not possible, a 409 Conflict status will be returned. - preconditions:: { - local __preconditionsMixin(preconditions) = __deleteOptionsMixin({ preconditions+: preconditions }), - mixinInstance(preconditions):: __preconditionsMixin(preconditions), - // Specifies the target ResourceVersion - withResourceVersion(resourceVersion):: self + __preconditionsMixin({ resourceVersion: resourceVersion }), - // Specifies the target UID. - withUid(uid):: self + __preconditionsMixin({ uid: uid }), - }, - preconditionsType:: hidden.meta.v1.preconditions, - // Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. - withPropagationPolicy(propagationPolicy):: self + __deleteOptionsMixin({ propagationPolicy: propagationPolicy }), - }, - deleteOptionsType:: hidden.meta.v1.deleteOptions, - // ObjectMeta describes the pod that is being evicted. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // PodDisruptionBudget is an object to define the max disruption that can be caused to a collection of pods - podDisruptionBudget:: { - local kind = { kind: 'PodDisruptionBudget' }, - new():: apiVersion + kind, - mixin:: { - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the PodDisruptionBudget. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // An eviction is allowed if at most "maxUnavailable" pods selected by "selector" are unavailable after the eviction, i.e. even in absence of the evicted pod. For example, one can prevent all voluntary evictions by specifying 0. This is a mutually exclusive setting with "minAvailable". - withMaxUnavailable(maxUnavailable):: self + __specMixin({ maxUnavailable: maxUnavailable }), - // An eviction is allowed if at least "minAvailable" pods selected by "selector" will still be available after the eviction, i.e. even in the absence of the evicted pod. So for example you can prevent all voluntary evictions by specifying "100%". - withMinAvailable(minAvailable):: self + __specMixin({ minAvailable: minAvailable }), - // Label query over pods whose evictions are managed by the disruption budget. - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - specType:: hidden.policy.v1beta1.podDisruptionBudgetSpec, - }, - }, - // PodSecurityPolicy governs the ability to make requests that affect the Security Context that will be applied to a pod and container. - podSecurityPolicy:: { - local kind = { kind: 'PodSecurityPolicy' }, - new():: apiVersion + kind, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // spec defines the policy enforced. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true. - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + __specMixin({ allowPrivilegeEscalation: allowPrivilegeEscalation }), - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDrivers(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then __specMixin({ allowedCSIDrivers: allowedCsiDrivers }) else __specMixin({ allowedCSIDrivers: [allowedCsiDrivers] }), - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDriversMixin(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then __specMixin({ allowedCSIDrivers+: allowedCsiDrivers }) else __specMixin({ allowedCSIDrivers+: [allowedCsiDrivers] }), - allowedCSIDriversType:: hidden.policy.v1beta1.allowedCsiDriver, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilities(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then __specMixin({ allowedCapabilities: allowedCapabilities }) else __specMixin({ allowedCapabilities: [allowedCapabilities] }), - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilitiesMixin(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then __specMixin({ allowedCapabilities+: allowedCapabilities }) else __specMixin({ allowedCapabilities+: [allowedCapabilities] }), - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumes(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then __specMixin({ allowedFlexVolumes: allowedFlexVolumes }) else __specMixin({ allowedFlexVolumes: [allowedFlexVolumes] }), - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumesMixin(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then __specMixin({ allowedFlexVolumes+: allowedFlexVolumes }) else __specMixin({ allowedFlexVolumes+: [allowedFlexVolumes] }), - allowedFlexVolumesType:: hidden.policy.v1beta1.allowedFlexVolume, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPaths(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then __specMixin({ allowedHostPaths: allowedHostPaths }) else __specMixin({ allowedHostPaths: [allowedHostPaths] }), - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPathsMixin(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then __specMixin({ allowedHostPaths+: allowedHostPaths }) else __specMixin({ allowedHostPaths+: [allowedHostPaths] }), - allowedHostPathsType:: hidden.policy.v1beta1.allowedHostPath, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypes(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then __specMixin({ allowedProcMountTypes: allowedProcMountTypes }) else __specMixin({ allowedProcMountTypes: [allowedProcMountTypes] }), - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypesMixin(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then __specMixin({ allowedProcMountTypes+: allowedProcMountTypes }) else __specMixin({ allowedProcMountTypes+: [allowedProcMountTypes] }), - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctls(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then __specMixin({ allowedUnsafeSysctls: allowedUnsafeSysctls }) else __specMixin({ allowedUnsafeSysctls: [allowedUnsafeSysctls] }), - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then __specMixin({ allowedUnsafeSysctls+: allowedUnsafeSysctls }) else __specMixin({ allowedUnsafeSysctls+: [allowedUnsafeSysctls] }), - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilities(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then __specMixin({ defaultAddCapabilities: defaultAddCapabilities }) else __specMixin({ defaultAddCapabilities: [defaultAddCapabilities] }), - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilitiesMixin(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then __specMixin({ defaultAddCapabilities+: defaultAddCapabilities }) else __specMixin({ defaultAddCapabilities+: [defaultAddCapabilities] }), - // defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process. - withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation):: self + __specMixin({ defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation }), - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctls(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then __specMixin({ forbiddenSysctls: forbiddenSysctls }) else __specMixin({ forbiddenSysctls: [forbiddenSysctls] }), - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctlsMixin(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then __specMixin({ forbiddenSysctls+: forbiddenSysctls }) else __specMixin({ forbiddenSysctls+: [forbiddenSysctls] }), - // fsGroup is the strategy that will dictate what fs group is used by the SecurityContext. - fsGroup:: { - local __fsGroupMixin(fsGroup) = __specMixin({ fsGroup+: fsGroup }), - mixinInstance(fsGroup):: __fsGroupMixin(fsGroup), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges: ranges }) else __fsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges+: ranges }) else __fsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + __fsGroupMixin({ rule: rule }), - }, - fsGroupType:: hidden.policy.v1beta1.fsGroupStrategyOptions, - // hostIPC determines if the policy allows the use of HostIPC in the pod spec. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // hostNetwork determines if the policy allows the use of HostNetwork in the pod spec. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // hostPID determines if the policy allows the use of HostPID in the pod spec. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPorts(hostPorts):: self + if std.type(hostPorts) == 'array' then __specMixin({ hostPorts: hostPorts }) else __specMixin({ hostPorts: [hostPorts] }), - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPortsMixin(hostPorts):: self + if std.type(hostPorts) == 'array' then __specMixin({ hostPorts+: hostPorts }) else __specMixin({ hostPorts+: [hostPorts] }), - hostPortsType:: hidden.policy.v1beta1.hostPortRange, - // privileged determines if a pod can request to be run as privileged. - withPrivileged(privileged):: self + __specMixin({ privileged: privileged }), - // readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to. - withReadOnlyRootFilesystem(readOnlyRootFilesystem):: self + __specMixin({ readOnlyRootFilesystem: readOnlyRootFilesystem }), - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilities(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then __specMixin({ requiredDropCapabilities: requiredDropCapabilities }) else __specMixin({ requiredDropCapabilities: [requiredDropCapabilities] }), - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilitiesMixin(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then __specMixin({ requiredDropCapabilities+: requiredDropCapabilities }) else __specMixin({ requiredDropCapabilities+: [requiredDropCapabilities] }), - // RunAsGroup is the strategy that will dictate the allowable RunAsGroup values that may be set. If this field is omitted, the pod's RunAsGroup can take any value. This field requires the RunAsGroup feature gate to be enabled. - runAsGroup:: { - local __runAsGroupMixin(runAsGroup) = __specMixin({ runAsGroup+: runAsGroup }), - mixinInstance(runAsGroup):: __runAsGroupMixin(runAsGroup), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges: ranges }) else __runAsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges+: ranges }) else __runAsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + __runAsGroupMixin({ rule: rule }), - }, - runAsGroupType:: hidden.policy.v1beta1.runAsGroupStrategyOptions, - // runAsUser is the strategy that will dictate the allowable RunAsUser values that may be set. - runAsUser:: { - local __runAsUserMixin(runAsUser) = __specMixin({ runAsUser+: runAsUser }), - mixinInstance(runAsUser):: __runAsUserMixin(runAsUser), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges: ranges }) else __runAsUserMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges+: ranges }) else __runAsUserMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + __runAsUserMixin({ rule: rule }), - }, - runAsUserType:: hidden.policy.v1beta1.runAsUserStrategyOptions, - // seLinux is the strategy that will dictate the allowable labels that may be set. - seLinux:: { - local __seLinuxMixin(seLinux) = __specMixin({ seLinux+: seLinux }), - mixinInstance(seLinux):: __seLinuxMixin(seLinux), - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + __seLinuxMixin({ rule: rule }), - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __seLinuxMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - seLinuxType:: hidden.policy.v1beta1.seLinuxStrategyOptions, - // supplementalGroups is the strategy that will dictate what supplemental groups are used by the SecurityContext. - supplementalGroups:: { - local __supplementalGroupsMixin(supplementalGroups) = __specMixin({ supplementalGroups+: supplementalGroups }), - mixinInstance(supplementalGroups):: __supplementalGroupsMixin(supplementalGroups), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges: ranges }) else __supplementalGroupsMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges+: ranges }) else __supplementalGroupsMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + __supplementalGroupsMixin({ rule: rule }), - }, - supplementalGroupsType:: hidden.policy.v1beta1.supplementalGroupsStrategyOptions, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - }, - specType:: hidden.policy.v1beta1.podSecurityPolicySpec, - }, - }, - }, - }, - rbac:: { - v1:: { - local apiVersion = { apiVersion: 'rbac.authorization.k8s.io/v1' }, - // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. - clusterRole:: { - local kind = { kind: 'ClusterRole' }, - new():: apiVersion + kind, - // Rules holds all the PolicyRules for this ClusterRole - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // Rules holds all the PolicyRules for this ClusterRole - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.rbac.v1.policyRule, - mixin:: { - // AggregationRule is an optional field that describes how to build the Rules for this ClusterRole. If AggregationRule is set, then the Rules are controller managed and direct changes to Rules will be stomped by the controller. - aggregationRule:: { - local __aggregationRuleMixin(aggregationRule) = { aggregationRule+: aggregationRule }, - mixinInstance(aggregationRule):: __aggregationRuleMixin(aggregationRule), - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectors(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then __aggregationRuleMixin({ clusterRoleSelectors: clusterRoleSelectors }) else __aggregationRuleMixin({ clusterRoleSelectors: [clusterRoleSelectors] }), - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectorsMixin(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then __aggregationRuleMixin({ clusterRoleSelectors+: clusterRoleSelectors }) else __aggregationRuleMixin({ clusterRoleSelectors+: [clusterRoleSelectors] }), - clusterRoleSelectorsType:: hidden.meta.v1.labelSelector, - }, - aggregationRuleType:: hidden.rbac.v1.aggregationRule, - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. - clusterRoleBinding:: { - local kind = { kind: 'ClusterRoleBinding' }, - new():: apiVersion + kind, - // Subjects holds references to the objects the role applies to. - withSubjects(subjects):: self + if std.type(subjects) == 'array' then { subjects: subjects } else { subjects: [subjects] }, - // Subjects holds references to the objects the role applies to. - withSubjectsMixin(subjects):: self + if std.type(subjects) == 'array' then { subjects+: subjects } else { subjects+: [subjects] }, - subjectsType:: hidden.rbac.v1.subject, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // RoleRef can only reference a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error. - roleRef:: { - local __roleRefMixin(roleRef) = { roleRef+: roleRef }, - mixinInstance(roleRef):: __roleRefMixin(roleRef), - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + __roleRefMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __roleRefMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __roleRefMixin({ name: name }), - }, - roleRefType:: hidden.rbac.v1.roleRef, - }, - }, - // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. - role:: { - local kind = { kind: 'Role' }, - new():: apiVersion + kind, - // Rules holds all the PolicyRules for this Role - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // Rules holds all the PolicyRules for this Role - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.rbac.v1.policyRule, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. - roleBinding:: { - local kind = { kind: 'RoleBinding' }, - new():: apiVersion + kind, - // Subjects holds references to the objects the role applies to. - withSubjects(subjects):: self + if std.type(subjects) == 'array' then { subjects: subjects } else { subjects: [subjects] }, - // Subjects holds references to the objects the role applies to. - withSubjectsMixin(subjects):: self + if std.type(subjects) == 'array' then { subjects+: subjects } else { subjects+: [subjects] }, - subjectsType:: hidden.rbac.v1.subject, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // RoleRef can reference a Role in the current namespace or a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error. - roleRef:: { - local __roleRefMixin(roleRef) = { roleRef+: roleRef }, - mixinInstance(roleRef):: __roleRefMixin(roleRef), - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + __roleRefMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __roleRefMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __roleRefMixin({ name: name }), - }, - roleRefType:: hidden.rbac.v1.roleRef, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'rbac.authorization.k8s.io/v1beta1' }, - // ClusterRole is a cluster level, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding or ClusterRoleBinding. - clusterRole:: { - local kind = { kind: 'ClusterRole' }, - new():: apiVersion + kind, - // Rules holds all the PolicyRules for this ClusterRole - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // Rules holds all the PolicyRules for this ClusterRole - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.rbac.v1beta1.policyRule, - mixin:: { - // AggregationRule is an optional field that describes how to build the Rules for this ClusterRole. If AggregationRule is set, then the Rules are controller managed and direct changes to Rules will be stomped by the controller. - aggregationRule:: { - local __aggregationRuleMixin(aggregationRule) = { aggregationRule+: aggregationRule }, - mixinInstance(aggregationRule):: __aggregationRuleMixin(aggregationRule), - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectors(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then __aggregationRuleMixin({ clusterRoleSelectors: clusterRoleSelectors }) else __aggregationRuleMixin({ clusterRoleSelectors: [clusterRoleSelectors] }), - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectorsMixin(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then __aggregationRuleMixin({ clusterRoleSelectors+: clusterRoleSelectors }) else __aggregationRuleMixin({ clusterRoleSelectors+: [clusterRoleSelectors] }), - clusterRoleSelectorsType:: hidden.meta.v1.labelSelector, - }, - aggregationRuleType:: hidden.rbac.v1beta1.aggregationRule, - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // ClusterRoleBinding references a ClusterRole, but not contain it. It can reference a ClusterRole in the global namespace, and adds who information via Subject. - clusterRoleBinding:: { - local kind = { kind: 'ClusterRoleBinding' }, - new():: apiVersion + kind, - // Subjects holds references to the objects the role applies to. - withSubjects(subjects):: self + if std.type(subjects) == 'array' then { subjects: subjects } else { subjects: [subjects] }, - // Subjects holds references to the objects the role applies to. - withSubjectsMixin(subjects):: self + if std.type(subjects) == 'array' then { subjects+: subjects } else { subjects+: [subjects] }, - subjectsType:: hidden.rbac.v1beta1.subject, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // RoleRef can only reference a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error. - roleRef:: { - local __roleRefMixin(roleRef) = { roleRef+: roleRef }, - mixinInstance(roleRef):: __roleRefMixin(roleRef), - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + __roleRefMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __roleRefMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __roleRefMixin({ name: name }), - }, - roleRefType:: hidden.rbac.v1beta1.roleRef, - }, - }, - // Role is a namespaced, logical grouping of PolicyRules that can be referenced as a unit by a RoleBinding. - role:: { - local kind = { kind: 'Role' }, - new():: apiVersion + kind, - // Rules holds all the PolicyRules for this Role - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // Rules holds all the PolicyRules for this Role - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.rbac.v1beta1.policyRule, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // RoleBinding references a role, but does not contain it. It can reference a Role in the same namespace or a ClusterRole in the global namespace. It adds who information via Subjects and namespace information by which namespace it exists in. RoleBindings in a given namespace only have effect in that namespace. - roleBinding:: { - local kind = { kind: 'RoleBinding' }, - new():: apiVersion + kind, - // Subjects holds references to the objects the role applies to. - withSubjects(subjects):: self + if std.type(subjects) == 'array' then { subjects: subjects } else { subjects: [subjects] }, - // Subjects holds references to the objects the role applies to. - withSubjectsMixin(subjects):: self + if std.type(subjects) == 'array' then { subjects+: subjects } else { subjects+: [subjects] }, - subjectsType:: hidden.rbac.v1beta1.subject, - mixin:: { - // Standard object's metadata. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // RoleRef can reference a Role in the current namespace or a ClusterRole in the global namespace. If the RoleRef cannot be resolved, the Authorizer must return an error. - roleRef:: { - local __roleRefMixin(roleRef) = { roleRef+: roleRef }, - mixinInstance(roleRef):: __roleRefMixin(roleRef), - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + __roleRefMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __roleRefMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __roleRefMixin({ name: name }), - }, - roleRefType:: hidden.rbac.v1beta1.roleRef, - }, - }, - }, - }, - scheduling:: { - v1:: { - local apiVersion = { apiVersion: 'scheduling.k8s.io/v1' }, - // PriorityClass defines mapping from a priority class name to the priority integer value. The value can be any valid integer. - priorityClass:: { - local kind = { kind: 'PriorityClass' }, - new():: apiVersion + kind, - // description is an arbitrary string that usually provides guidelines on when this priority class should be used. - withDescription(description):: self + { description: description }, - // globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority. - withGlobalDefault(globalDefault):: self + { globalDefault: globalDefault }, - // The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec. - withValue(value):: self + { value: value }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'scheduling.k8s.io/v1beta1' }, - // DEPRECATED - This group version of PriorityClass is deprecated by scheduling.k8s.io/v1/PriorityClass. PriorityClass defines mapping from a priority class name to the priority integer value. The value can be any valid integer. - priorityClass:: { - local kind = { kind: 'PriorityClass' }, - new():: apiVersion + kind, - // description is an arbitrary string that usually provides guidelines on when this priority class should be used. - withDescription(description):: self + { description: description }, - // globalDefault specifies whether this PriorityClass should be considered as the default priority for pods that do not have any priority class. Only one PriorityClass can be marked as `globalDefault`. However, if more than one PriorityClasses exists with their `globalDefault` field set to true, the smallest value of such global default PriorityClasses will be used as the default priority. - withGlobalDefault(globalDefault):: self + { globalDefault: globalDefault }, - // The value of this priority class. This is the actual priority that pods receive when they have the name of this class in their pod spec. - withValue(value):: self + { value: value }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - }, - }, - storage:: { - v1:: { - local apiVersion = { apiVersion: 'storage.k8s.io/v1' }, - // StorageClass describes the parameters for a class of storage for which PersistentVolumes can be dynamically provisioned. - // - // StorageClasses are non-namespaced; the name of the storage class according to etcd is in ObjectMeta.Name. - storageClass:: { - local kind = { kind: 'StorageClass' }, - new():: apiVersion + kind, - // AllowVolumeExpansion shows whether the storage class allow volume expand - withAllowVolumeExpansion(allowVolumeExpansion):: self + { allowVolumeExpansion: allowVolumeExpansion }, - // Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature. - withAllowedTopologies(allowedTopologies):: self + if std.type(allowedTopologies) == 'array' then { allowedTopologies: allowedTopologies } else { allowedTopologies: [allowedTopologies] }, - // Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature. - withAllowedTopologiesMixin(allowedTopologies):: self + if std.type(allowedTopologies) == 'array' then { allowedTopologies+: allowedTopologies } else { allowedTopologies+: [allowedTopologies] }, - allowedTopologiesType:: hidden.core.v1.topologySelectorTerm, - // Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid. - withMountOptions(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions: mountOptions } else { mountOptions: [mountOptions] }, - // Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid. - withMountOptionsMixin(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions+: mountOptions } else { mountOptions+: [mountOptions] }, - // Parameters holds the parameters for the provisioner that should create volumes of this storage class. - withParameters(parameters):: self + { parameters: parameters }, - // Parameters holds the parameters for the provisioner that should create volumes of this storage class. - withParametersMixin(parameters):: self + { parameters+: parameters }, - // Provisioner indicates the type of the provisioner. - withProvisioner(provisioner):: self + { provisioner: provisioner }, - // Dynamically provisioned PersistentVolumes of this storage class are created with this reclaimPolicy. Defaults to Delete. - withReclaimPolicy(reclaimPolicy):: self + { reclaimPolicy: reclaimPolicy }, - // VolumeBindingMode indicates how PersistentVolumeClaims should be provisioned and bound. When unset, VolumeBindingImmediate is used. This field is only honored by servers that enable the VolumeScheduling feature. - withVolumeBindingMode(volumeBindingMode):: self + { volumeBindingMode: volumeBindingMode }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node. - // - // VolumeAttachment objects are non-namespaced. - volumeAttachment:: { - local kind = { kind: 'VolumeAttachment' }, - new():: apiVersion + kind, - mixin:: { - // Standard object metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired attach/detach volume behavior. Populated by the Kubernetes system. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName(). - withAttacher(attacher):: self + __specMixin({ attacher: attacher }), - // The node that the volume should be attached to. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // Source represents the volume that should be attached. - source:: { - local __sourceMixin(source) = __specMixin({ source+: source }), - mixinInstance(source):: __sourceMixin(source), - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + __sourceMixin({ persistentVolumeName: persistentVolumeName }), - }, - sourceType:: hidden.storage.v1.volumeAttachmentSource, - }, - specType:: hidden.storage.v1.volumeAttachmentSpec, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'storage.k8s.io/v1beta1' }, - // CSIDriver captures information about a Container Storage Interface (CSI) volume driver deployed on the cluster. CSI drivers do not need to create the CSIDriver object directly. Instead they may use the cluster-driver-registrar sidecar container. When deployed with a CSI driver it automatically creates a CSIDriver object representing the driver. Kubernetes attach detach controller uses this object to determine whether attach is required. Kubelet uses this object to determine whether pod information needs to be passed on mount. CSIDriver objects are non-namespaced. - csiDriver:: { - local kind = { kind: 'CSIDriver' }, - new():: apiVersion + kind, - mixin:: { - // Standard object metadata. metadata.Name indicates the name of the CSI driver that this object refers to; it MUST be the same name returned by the CSI GetPluginName() call for that driver. The driver name must be 63 characters or less, beginning and ending with an alphanumeric character ([a-z0-9A-Z]) with dashes (-), dots (.), and alphanumerics between. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the CSI Driver. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called. - withAttachRequired(attachRequired):: self + __specMixin({ attachRequired: attachRequired }), - // If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. "csi.storage.k8s.io/pod.name": pod.Name "csi.storage.k8s.io/pod.namespace": pod.Namespace "csi.storage.k8s.io/pod.uid": string(pod.UID) - withPodInfoOnMount(podInfoOnMount):: self + __specMixin({ podInfoOnMount: podInfoOnMount }), - }, - specType:: hidden.storage.v1beta1.csiDriverSpec, - }, - }, - // CSINode holds information about all CSI drivers installed on a node. CSI drivers do not need to create the CSINode object directly. As long as they use the node-driver-registrar sidecar container, the kubelet will automatically populate the CSINode object for the CSI driver as part of kubelet plugin registration. CSINode has the same name as a node. If the object is missing, it means either there are no CSI Drivers available on the node, or the Kubelet version is low enough that it doesn't create this object. CSINode has an OwnerReference that points to the corresponding node object. - csiNode:: { - local kind = { kind: 'CSINode' }, - new():: apiVersion + kind, - mixin:: { - // metadata.name must be the Kubernetes node name. - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // spec is the specification of CSINode - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty. - withDrivers(drivers):: self + if std.type(drivers) == 'array' then __specMixin({ drivers: drivers }) else __specMixin({ drivers: [drivers] }), - // drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty. - withDriversMixin(drivers):: self + if std.type(drivers) == 'array' then __specMixin({ drivers+: drivers }) else __specMixin({ drivers+: [drivers] }), - driversType:: hidden.storage.v1beta1.csiNodeDriver, - }, - specType:: hidden.storage.v1beta1.csiNodeSpec, - }, - }, - // StorageClass describes the parameters for a class of storage for which PersistentVolumes can be dynamically provisioned. - // - // StorageClasses are non-namespaced; the name of the storage class according to etcd is in ObjectMeta.Name. - storageClass:: { - local kind = { kind: 'StorageClass' }, - new():: apiVersion + kind, - // AllowVolumeExpansion shows whether the storage class allow volume expand - withAllowVolumeExpansion(allowVolumeExpansion):: self + { allowVolumeExpansion: allowVolumeExpansion }, - // Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature. - withAllowedTopologies(allowedTopologies):: self + if std.type(allowedTopologies) == 'array' then { allowedTopologies: allowedTopologies } else { allowedTopologies: [allowedTopologies] }, - // Restrict the node topologies where volumes can be dynamically provisioned. Each volume plugin defines its own supported topology specifications. An empty TopologySelectorTerm list means there is no topology restriction. This field is only honored by servers that enable the VolumeScheduling feature. - withAllowedTopologiesMixin(allowedTopologies):: self + if std.type(allowedTopologies) == 'array' then { allowedTopologies+: allowedTopologies } else { allowedTopologies+: [allowedTopologies] }, - allowedTopologiesType:: hidden.core.v1.topologySelectorTerm, - // Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid. - withMountOptions(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions: mountOptions } else { mountOptions: [mountOptions] }, - // Dynamically provisioned PersistentVolumes of this storage class are created with these mountOptions, e.g. ["ro", "soft"]. Not validated - mount of the PVs will simply fail if one is invalid. - withMountOptionsMixin(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions+: mountOptions } else { mountOptions+: [mountOptions] }, - // Parameters holds the parameters for the provisioner that should create volumes of this storage class. - withParameters(parameters):: self + { parameters: parameters }, - // Parameters holds the parameters for the provisioner that should create volumes of this storage class. - withParametersMixin(parameters):: self + { parameters+: parameters }, - // Provisioner indicates the type of the provisioner. - withProvisioner(provisioner):: self + { provisioner: provisioner }, - // Dynamically provisioned PersistentVolumes of this storage class are created with this reclaimPolicy. Defaults to Delete. - withReclaimPolicy(reclaimPolicy):: self + { reclaimPolicy: reclaimPolicy }, - // VolumeBindingMode indicates how PersistentVolumeClaims should be provisioned and bound. When unset, VolumeBindingImmediate is used. This field is only honored by servers that enable the VolumeScheduling feature. - withVolumeBindingMode(volumeBindingMode):: self + { volumeBindingMode: volumeBindingMode }, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // VolumeAttachment captures the intent to attach or detach the specified volume to/from the specified node. - // - // VolumeAttachment objects are non-namespaced. - volumeAttachment:: { - local kind = { kind: 'VolumeAttachment' }, - new():: apiVersion + kind, - mixin:: { - // Standard object metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired attach/detach volume behavior. Populated by the Kubernetes system. - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName(). - withAttacher(attacher):: self + __specMixin({ attacher: attacher }), - // The node that the volume should be attached to. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // Source represents the volume that should be attached. - source:: { - local __sourceMixin(source) = __specMixin({ source+: source }), - mixinInstance(source):: __sourceMixin(source), - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + __sourceMixin({ persistentVolumeName: persistentVolumeName }), - }, - sourceType:: hidden.storage.v1beta1.volumeAttachmentSource, - }, - specType:: hidden.storage.v1beta1.volumeAttachmentSpec, - }, - }, - }, - }, - local hidden = { - admissionregistration:: { - v1beta1:: { - local apiVersion = { apiVersion: 'admissionregistration/v1beta1' }, - // MutatingWebhookConfigurationList is a list of MutatingWebhookConfiguration. - mutatingWebhookConfigurationList:: { - new():: {}, - // List of MutatingWebhookConfiguration. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of MutatingWebhookConfiguration. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.admissionregistration.v1beta1.mutatingWebhookConfiguration, - mixin:: {}, - }, - // RuleWithOperations is a tuple of Operations and Resources. It is recommended to make sure that all the tuple expansions are valid. - ruleWithOperations:: { - new():: {}, - // APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required. - withApiGroups(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups: apiGroups } else { apiGroups: [apiGroups] }, - // APIGroups is the API groups the resources belong to. '*' is all groups. If '*' is present, the length of the slice must be one. Required. - withApiGroupsMixin(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups+: apiGroups } else { apiGroups+: [apiGroups] }, - // APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required. - withApiVersions(apiVersions):: self + if std.type(apiVersions) == 'array' then { apiVersions: apiVersions } else { apiVersions: [apiVersions] }, - // APIVersions is the API versions the resources belong to. '*' is all versions. If '*' is present, the length of the slice must be one. Required. - withApiVersionsMixin(apiVersions):: self + if std.type(apiVersions) == 'array' then { apiVersions+: apiVersions } else { apiVersions+: [apiVersions] }, - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * for all operations. If '*' is present, the length of the slice must be one. Required. - withOperations(operations):: self + if std.type(operations) == 'array' then { operations: operations } else { operations: [operations] }, - // Operations is the operations the admission hook cares about - CREATE, UPDATE, or * for all operations. If '*' is present, the length of the slice must be one. Required. - withOperationsMixin(operations):: self + if std.type(operations) == 'array' then { operations+: operations } else { operations+: [operations] }, - // Resources is a list of resources this rule applies to. - // - // For example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources. - // - // If wildcard is present, the validation rule will ensure resources do not overlap with each other. - // - // Depending on the enclosing object, subresources might not be allowed. Required. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // Resources is a list of resources this rule applies to. - // - // For example: 'pods' means pods. 'pods/log' means the log subresource of pods. '*' means all resources, but not subresources. 'pods/*' means all subresources of pods. '*/scale' means all scale subresources. '*/*' means all resources and their subresources. - // - // If wildcard is present, the validation rule will ensure resources do not overlap with each other. - // - // Depending on the enclosing object, subresources might not be allowed. Required. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - // scope specifies the scope of this rule. Valid values are "Cluster", "Namespaced", and "*" "Cluster" means that only cluster-scoped resources will match this rule. Namespace API objects are cluster-scoped. "Namespaced" means that only namespaced resources will match this rule. "*" means that there are no scope restrictions. Subresources match the scope of their parent resource. Default is "*". - withScope(scope):: self + { scope: scope }, - mixin:: {}, - }, - // ServiceReference holds a reference to Service.legacy.k8s.io - serviceReference:: { - new():: {}, - // `name` is the name of the service. Required - withName(name):: self + { name: name }, - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + { namespace: namespace }, - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + { path: path }, - mixin:: {}, - }, - // ValidatingWebhookConfigurationList is a list of ValidatingWebhookConfiguration. - validatingWebhookConfigurationList:: { - new():: {}, - // List of ValidatingWebhookConfiguration. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ValidatingWebhookConfiguration. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.admissionregistration.v1beta1.validatingWebhookConfiguration, - mixin:: {}, - }, - // Webhook describes an admission webhook and the resources and operations it applies to. - webhook:: { - new():: {}, - // AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`. - withAdmissionReviewVersions(admissionReviewVersions):: self + if std.type(admissionReviewVersions) == 'array' then { admissionReviewVersions: admissionReviewVersions } else { admissionReviewVersions: [admissionReviewVersions] }, - // AdmissionReviewVersions is an ordered list of preferred `AdmissionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, validation will fail for this object. If a persisted webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail and be subject to the failure policy. Default to `['v1beta1']`. - withAdmissionReviewVersionsMixin(admissionReviewVersions):: self + if std.type(admissionReviewVersions) == 'array' then { admissionReviewVersions+: admissionReviewVersions } else { admissionReviewVersions+: [admissionReviewVersions] }, - // FailurePolicy defines how unrecognized errors from the admission endpoint are handled - allowed values are Ignore or Fail. Defaults to Ignore. - withFailurePolicy(failurePolicy):: self + { failurePolicy: failurePolicy }, - // The name of the admission webhook. Name should be fully qualified, e.g., imagepolicy.kubernetes.io, where "imagepolicy" is the name of the webhook, and kubernetes.io is the name of the organization. Required. - withName(name):: self + { name: name }, - // Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects. - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // Rules describes what operations on what resources/subresources the webhook cares about. The webhook cares about an operation if it matches _any_ Rule. However, in order to prevent ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks from putting the cluster in a state which cannot be recovered from without completely disabling the plugin, ValidatingAdmissionWebhooks and MutatingAdmissionWebhooks are never called on admission requests for ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects. - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.admissionregistration.v1beta1.ruleWithOperations, - // SideEffects states whether this webhookk has side effects. Acceptable values are: Unknown, None, Some, NoneOnDryRun Webhooks with side effects MUST implement a reconciliation system, since a request may be rejected by a future step in the admission change and the side effects therefore need to be undone. Requests with the dryRun attribute will be auto-rejected if they match a webhook with sideEffects == Unknown or Some. Defaults to Unknown. - withSideEffects(sideEffects):: self + { sideEffects: sideEffects }, - // TimeoutSeconds specifies the timeout for this webhook. After the timeout passes, the webhook call will be ignored or the API call will fail based on the failure policy. The timeout value must be between 1 and 30 seconds. Default to 30 seconds. - withTimeoutSeconds(timeoutSeconds):: self + { timeoutSeconds: timeoutSeconds }, - mixin:: { - // ClientConfig defines how to communicate with the hook. Required - clientConfig:: { - local __clientConfigMixin(clientConfig) = { clientConfig+: clientConfig }, - mixinInstance(clientConfig):: __clientConfigMixin(clientConfig), - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __clientConfigMixin({ caBundle: caBundle }), - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = __clientConfigMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.admissionregistration.v1beta1.serviceReference, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + __clientConfigMixin({ url: url }), - }, - clientConfigType:: hidden.admissionregistration.v1beta1.webhookClientConfig, - // NamespaceSelector decides whether to run the webhook on an object based on whether the namespace for that object matches the selector. If the object itself is a namespace, the matching is performed on object.metadata.labels. If the object is another cluster scoped resource, it never skips the webhook. - // - // For example, to run the webhook on any objects whose namespace is not associated with "runlevel" of "0" or "1"; you will set the selector as follows: "namespaceSelector": { - // "matchExpressions": [ - // { - // "key": "runlevel", - // "operator": "NotIn", - // "values": [ - // "0", - // "1" - // ] - // } - // ] - // } - // - // If instead you want to only run the webhook on any objects whose namespace is associated with the "environment" of "prod" or "staging"; you will set the selector as follows: "namespaceSelector": { - // "matchExpressions": [ - // { - // "key": "environment", - // "operator": "In", - // "values": [ - // "prod", - // "staging" - // ] - // } - // ] - // } - // - // See https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/ for more examples of label selectors. - // - // Default to the empty LabelSelector, which matches everything. - namespaceSelector:: { - local __namespaceSelectorMixin(namespaceSelector) = { namespaceSelector+: namespaceSelector }, - mixinInstance(namespaceSelector):: __namespaceSelectorMixin(namespaceSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions+: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __namespaceSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __namespaceSelectorMixin({ matchLabels+: matchLabels }), - }, - namespaceSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // WebhookClientConfig contains the information to make a TLS connection with the webhook - webhookClientConfig:: { - new():: {}, - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + { caBundle: caBundle }, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + { url: url }, - mixin:: { - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = { service+: service }, - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.admissionregistration.v1beta1.serviceReference, - }, - }, - }, - }, - apiextensions:: { - v1beta1:: { - local apiVersion = { apiVersion: 'apiextensions/v1beta1' }, - // CustomResourceColumnDefinition specifies a column for server side printing. - customResourceColumnDefinition:: { - new():: {}, - // JSONPath is a simple JSON path, i.e. with array notation. - withJsonPath(jsonPath):: self + { JSONPath: jsonPath }, - // description is a human readable description of this column. - withDescription(description):: self + { description: description }, - // format is an optional OpenAPI type definition for this column. The 'name' format is applied to the primary identifier column to assist in clients identifying column is the resource name. See https://github.com/OAI/OpenAPI-Specification/blob/master/versions/2.0.md#data-types for more. - withFormat(format):: self + { format: format }, - // name is a human readable name for the column. - withName(name):: self + { name: name }, - // priority is an integer defining the relative importance of this column compared to others. Lower numbers are considered higher priority. Columns that may be omitted in limited space scenarios should be given a higher priority. - withPriority(priority):: self + { priority: priority }, - // type is an OpenAPI type definition for this column. See https://github.com/OAI/OpenAPI-Specification/blob/master/versions/2.0.md#data-types for more. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // CustomResourceConversion describes how to convert different versions of a CR. - customResourceConversion:: { - new():: {}, - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersions(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then { conversionReviewVersions: conversionReviewVersions } else { conversionReviewVersions: [conversionReviewVersions] }, - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersionsMixin(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then { conversionReviewVersions+: conversionReviewVersions } else { conversionReviewVersions+: [conversionReviewVersions] }, - // `strategy` specifies the conversion strategy. Allowed values are: - `None`: The converter only change the apiVersion and would not touch any other field in the CR. - `Webhook`: API Server will call to an external webhook to do the conversion. Additional information is needed for this option. - withStrategy(strategy):: self + { strategy: strategy }, - mixin:: { - // `webhookClientConfig` is the instructions for how to call the webhook if strategy is `Webhook`. This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. - webhookClientConfig:: { - local __webhookClientConfigMixin(webhookClientConfig) = { webhookClientConfig+: webhookClientConfig }, - mixinInstance(webhookClientConfig):: __webhookClientConfigMixin(webhookClientConfig), - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __webhookClientConfigMixin({ caBundle: caBundle }), - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = __webhookClientConfigMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.apiextensions.v1beta1.serviceReference, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + __webhookClientConfigMixin({ url: url }), - }, - webhookClientConfigType:: hidden.apiextensions.v1beta1.webhookClientConfig, - }, - }, - // CustomResourceDefinitionCondition contains details for the current condition of this pod. - customResourceDefinitionCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human-readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // Unique, one-word, CamelCase reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status is the status of the condition. Can be True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type is the type of the condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // CustomResourceDefinitionList is a list of CustomResourceDefinition objects. - customResourceDefinitionList:: { - new():: {}, - // Items individual CustomResourceDefinitions - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items individual CustomResourceDefinitions - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apiextensions.v1beta1.customResourceDefinition, - mixin:: {}, - }, - // CustomResourceDefinitionNames indicates the names to serve this CustomResourceDefinition - customResourceDefinitionNames:: { - new():: {}, - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategories(categories):: self + if std.type(categories) == 'array' then { categories: categories } else { categories: [categories] }, - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategoriesMixin(categories):: self + if std.type(categories) == 'array' then { categories+: categories } else { categories+: [categories] }, - // Kind is the serialized kind of the resource. It is normally CamelCase and singular. - withKind(kind):: self + { kind: kind }, - // ListKind is the serialized kind of the list for this resource. Defaults to List. - withListKind(listKind):: self + { listKind: listKind }, - // Plural is the plural name of the resource to serve. It must match the name of the CustomResourceDefinition-registration too: plural.group and it must be all lowercase. - withPlural(plural):: self + { plural: plural }, - // ShortNames are short names for the resource. It must be all lowercase. - withShortNames(shortNames):: self + if std.type(shortNames) == 'array' then { shortNames: shortNames } else { shortNames: [shortNames] }, - // ShortNames are short names for the resource. It must be all lowercase. - withShortNamesMixin(shortNames):: self + if std.type(shortNames) == 'array' then { shortNames+: shortNames } else { shortNames+: [shortNames] }, - // Singular is the singular name of the resource. It must be all lowercase Defaults to lowercased - withSingular(singular):: self + { singular: singular }, - mixin:: {}, - }, - // CustomResourceDefinitionSpec describes how a user wants their resource to appear - customResourceDefinitionSpec:: { - new():: {}, - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Optional, the global columns for all versions. Top-level and per-version columns are mutually exclusive. - withAdditionalPrinterColumns(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then { additionalPrinterColumns: additionalPrinterColumns } else { additionalPrinterColumns: [additionalPrinterColumns] }, - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Optional, the global columns for all versions. Top-level and per-version columns are mutually exclusive. - withAdditionalPrinterColumnsMixin(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then { additionalPrinterColumns+: additionalPrinterColumns } else { additionalPrinterColumns+: [additionalPrinterColumns] }, - additionalPrinterColumnsType:: hidden.apiextensions.v1beta1.customResourceColumnDefinition, - // Group is the group this resource belongs in - withGroup(group):: self + { group: group }, - // Scope indicates whether this resource is cluster or namespace scoped. Default is namespaced - withScope(scope):: self + { scope: scope }, - // Version is the version this resource belongs in Should be always first item in Versions field if provided. Optional, but at least one of Version or Versions must be set. Deprecated: Please use `Versions`. - withVersion(version):: self + { version: version }, - // Versions is the list of all supported versions for this resource. If Version field is provided, this field is optional. Validation: All versions must use the same validation schema for now. i.e., top level Validation field is applied to all of these versions. Order: The version name will be used to compute the order. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersions(versions):: self + if std.type(versions) == 'array' then { versions: versions } else { versions: [versions] }, - // Versions is the list of all supported versions for this resource. If Version field is provided, this field is optional. Validation: All versions must use the same validation schema for now. i.e., top level Validation field is applied to all of these versions. Order: The version name will be used to compute the order. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionsMixin(versions):: self + if std.type(versions) == 'array' then { versions+: versions } else { versions+: [versions] }, - versionsType:: hidden.apiextensions.v1beta1.customResourceDefinitionVersion, - mixin:: { - // `conversion` defines conversion settings for the CRD. - conversion:: { - local __conversionMixin(conversion) = { conversion+: conversion }, - mixinInstance(conversion):: __conversionMixin(conversion), - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersions(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then __conversionMixin({ conversionReviewVersions: conversionReviewVersions }) else __conversionMixin({ conversionReviewVersions: [conversionReviewVersions] }), - // ConversionReviewVersions is an ordered list of preferred `ConversionReview` versions the Webhook expects. API server will try to use first version in the list which it supports. If none of the versions specified in this list supported by API server, conversion will fail for this object. If a persisted Webhook configuration specifies allowed versions and does not include any versions known to the API Server, calls to the webhook will fail. Default to `['v1beta1']`. - withConversionReviewVersionsMixin(conversionReviewVersions):: self + if std.type(conversionReviewVersions) == 'array' then __conversionMixin({ conversionReviewVersions+: conversionReviewVersions }) else __conversionMixin({ conversionReviewVersions+: [conversionReviewVersions] }), - // `strategy` specifies the conversion strategy. Allowed values are: - `None`: The converter only change the apiVersion and would not touch any other field in the CR. - `Webhook`: API Server will call to an external webhook to do the conversion. Additional information is needed for this option. - withStrategy(strategy):: self + __conversionMixin({ strategy: strategy }), - // `webhookClientConfig` is the instructions for how to call the webhook if strategy is `Webhook`. This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. - webhookClientConfig:: { - local __webhookClientConfigMixin(webhookClientConfig) = __conversionMixin({ webhookClientConfig+: webhookClientConfig }), - mixinInstance(webhookClientConfig):: __webhookClientConfigMixin(webhookClientConfig), - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + __webhookClientConfigMixin({ caBundle: caBundle }), - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = __webhookClientConfigMixin({ service+: service }), - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.apiextensions.v1beta1.serviceReference, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + __webhookClientConfigMixin({ url: url }), - }, - webhookClientConfigType:: hidden.apiextensions.v1beta1.webhookClientConfig, - }, - conversionType:: hidden.apiextensions.v1beta1.customResourceConversion, - // Names are the names used to describe this custom resource - names:: { - local __namesMixin(names) = { names+: names }, - mixinInstance(names):: __namesMixin(names), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategories(categories):: self + if std.type(categories) == 'array' then __namesMixin({ categories: categories }) else __namesMixin({ categories: [categories] }), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategoriesMixin(categories):: self + if std.type(categories) == 'array' then __namesMixin({ categories+: categories }) else __namesMixin({ categories+: [categories] }), - // Kind is the serialized kind of the resource. It is normally CamelCase and singular. - withKind(kind):: self + __namesMixin({ kind: kind }), - // ListKind is the serialized kind of the list for this resource. Defaults to List. - withListKind(listKind):: self + __namesMixin({ listKind: listKind }), - // Plural is the plural name of the resource to serve. It must match the name of the CustomResourceDefinition-registration too: plural.group and it must be all lowercase. - withPlural(plural):: self + __namesMixin({ plural: plural }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNames(shortNames):: self + if std.type(shortNames) == 'array' then __namesMixin({ shortNames: shortNames }) else __namesMixin({ shortNames: [shortNames] }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNamesMixin(shortNames):: self + if std.type(shortNames) == 'array' then __namesMixin({ shortNames+: shortNames }) else __namesMixin({ shortNames+: [shortNames] }), - // Singular is the singular name of the resource. It must be all lowercase Defaults to lowercased - withSingular(singular):: self + __namesMixin({ singular: singular }), - }, - namesType:: hidden.apiextensions.v1beta1.customResourceDefinitionNames, - // Subresources describes the subresources for CustomResource Optional, the global subresources for all versions. Top-level and per-version subresources are mutually exclusive. - subresources:: { - local __subresourcesMixin(subresources) = { subresources+: subresources }, - mixinInstance(subresources):: __subresourcesMixin(subresources), - // Scale denotes the scale subresource for CustomResources - scale:: { - local __scaleMixin(scale) = __subresourcesMixin({ scale+: scale }), - mixinInstance(scale):: __scaleMixin(scale), - // LabelSelectorPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Selector. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. Must be set to work with HPA. If there is no value under the given path in the CustomResource, the status label selector value in the /scale subresource will default to the empty string. - withLabelSelectorPath(labelSelectorPath):: self + __scaleMixin({ labelSelectorPath: labelSelectorPath }), - // SpecReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Spec.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .spec. If there is no value under the given path in the CustomResource, the /scale subresource will return an error on GET. - withSpecReplicasPath(specReplicasPath):: self + __scaleMixin({ specReplicasPath: specReplicasPath }), - // StatusReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. If there is no value under the given path in the CustomResource, the status replica value in the /scale subresource will default to 0. - withStatusReplicasPath(statusReplicasPath):: self + __scaleMixin({ statusReplicasPath: statusReplicasPath }), - }, - scaleType:: hidden.apiextensions.v1beta1.customResourceSubresourceScale, - }, - subresourcesType:: hidden.apiextensions.v1beta1.customResourceSubresources, - // Validation describes the validation methods for CustomResources Optional, the global validation schema for all versions. Top-level and per-version schemas are mutually exclusive. - validation:: { - local __validationMixin(validation) = { validation+: validation }, - mixinInstance(validation):: __validationMixin(validation), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3Schema(openApiV3Schema):: self + __validationMixin({ openAPIV3Schema: openApiV3Schema }), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3SchemaMixin(openApiV3Schema):: self + __validationMixin({ openAPIV3Schema+: openApiV3Schema }), - openAPIV3SchemaType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - }, - validationType:: hidden.apiextensions.v1beta1.customResourceValidation, - }, - }, - // CustomResourceDefinitionStatus indicates the state of the CustomResourceDefinition - customResourceDefinitionStatus:: { - new():: {}, - // Conditions indicate state for particular aspects of a CustomResourceDefinition - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Conditions indicate state for particular aspects of a CustomResourceDefinition - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apiextensions.v1beta1.customResourceDefinitionCondition, - // StoredVersions are all versions of CustomResources that were ever persisted. Tracking these versions allows a migration path for stored versions in etcd. The field is mutable so the migration controller can first finish a migration to another version (i.e. that no old objects are left in the storage), and then remove the rest of the versions from this list. None of the versions in this list can be removed from the spec.Versions field. - withStoredVersions(storedVersions):: self + if std.type(storedVersions) == 'array' then { storedVersions: storedVersions } else { storedVersions: [storedVersions] }, - // StoredVersions are all versions of CustomResources that were ever persisted. Tracking these versions allows a migration path for stored versions in etcd. The field is mutable so the migration controller can first finish a migration to another version (i.e. that no old objects are left in the storage), and then remove the rest of the versions from this list. None of the versions in this list can be removed from the spec.Versions field. - withStoredVersionsMixin(storedVersions):: self + if std.type(storedVersions) == 'array' then { storedVersions+: storedVersions } else { storedVersions+: [storedVersions] }, - mixin:: { - // AcceptedNames are the names that are actually being used to serve discovery They may be different than the names in spec. - acceptedNames:: { - local __acceptedNamesMixin(acceptedNames) = { acceptedNames+: acceptedNames }, - mixinInstance(acceptedNames):: __acceptedNamesMixin(acceptedNames), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategories(categories):: self + if std.type(categories) == 'array' then __acceptedNamesMixin({ categories: categories }) else __acceptedNamesMixin({ categories: [categories] }), - // Categories is a list of grouped resources custom resources belong to (e.g. 'all') - withCategoriesMixin(categories):: self + if std.type(categories) == 'array' then __acceptedNamesMixin({ categories+: categories }) else __acceptedNamesMixin({ categories+: [categories] }), - // Kind is the serialized kind of the resource. It is normally CamelCase and singular. - withKind(kind):: self + __acceptedNamesMixin({ kind: kind }), - // ListKind is the serialized kind of the list for this resource. Defaults to List. - withListKind(listKind):: self + __acceptedNamesMixin({ listKind: listKind }), - // Plural is the plural name of the resource to serve. It must match the name of the CustomResourceDefinition-registration too: plural.group and it must be all lowercase. - withPlural(plural):: self + __acceptedNamesMixin({ plural: plural }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNames(shortNames):: self + if std.type(shortNames) == 'array' then __acceptedNamesMixin({ shortNames: shortNames }) else __acceptedNamesMixin({ shortNames: [shortNames] }), - // ShortNames are short names for the resource. It must be all lowercase. - withShortNamesMixin(shortNames):: self + if std.type(shortNames) == 'array' then __acceptedNamesMixin({ shortNames+: shortNames }) else __acceptedNamesMixin({ shortNames+: [shortNames] }), - // Singular is the singular name of the resource. It must be all lowercase Defaults to lowercased - withSingular(singular):: self + __acceptedNamesMixin({ singular: singular }), - }, - acceptedNamesType:: hidden.apiextensions.v1beta1.customResourceDefinitionNames, - }, - }, - // CustomResourceDefinitionVersion describes a version for CRD. - customResourceDefinitionVersion:: { - new():: {}, - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Top-level and per-version columns are mutually exclusive. Per-version columns must not all be set to identical values (top-level columns should be used instead) This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. NOTE: CRDs created prior to 1.13 populated the top-level additionalPrinterColumns field by default. To apply an update that changes to per-version additionalPrinterColumns, the top-level additionalPrinterColumns field must be explicitly set to null - withAdditionalPrinterColumns(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then { additionalPrinterColumns: additionalPrinterColumns } else { additionalPrinterColumns: [additionalPrinterColumns] }, - // AdditionalPrinterColumns are additional columns shown e.g. in kubectl next to the name. Defaults to a created-at column. Top-level and per-version columns are mutually exclusive. Per-version columns must not all be set to identical values (top-level columns should be used instead) This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. NOTE: CRDs created prior to 1.13 populated the top-level additionalPrinterColumns field by default. To apply an update that changes to per-version additionalPrinterColumns, the top-level additionalPrinterColumns field must be explicitly set to null - withAdditionalPrinterColumnsMixin(additionalPrinterColumns):: self + if std.type(additionalPrinterColumns) == 'array' then { additionalPrinterColumns+: additionalPrinterColumns } else { additionalPrinterColumns+: [additionalPrinterColumns] }, - additionalPrinterColumnsType:: hidden.apiextensions.v1beta1.customResourceColumnDefinition, - // Name is the version name, e.g. “v1”, “v2beta1”, etc. - withName(name):: self + { name: name }, - // Served is a flag enabling/disabling this version from being served via REST APIs - withServed(served):: self + { served: served }, - // Storage flags the version as storage version. There must be exactly one flagged as storage version. - withStorage(storage):: self + { storage: storage }, - mixin:: { - // Schema describes the schema for CustomResource used in validation, pruning, and defaulting. Top-level and per-version schemas are mutually exclusive. Per-version schemas must not all be set to identical values (top-level validation schema should be used instead) This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. - schema:: { - local __schemaMixin(schema) = { schema+: schema }, - mixinInstance(schema):: __schemaMixin(schema), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3Schema(openApiV3Schema):: self + __schemaMixin({ openAPIV3Schema: openApiV3Schema }), - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3SchemaMixin(openApiV3Schema):: self + __schemaMixin({ openAPIV3Schema+: openApiV3Schema }), - openAPIV3SchemaType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - }, - schemaType:: hidden.apiextensions.v1beta1.customResourceValidation, - // Subresources describes the subresources for CustomResource Top-level and per-version subresources are mutually exclusive. Per-version subresources must not all be set to identical values (top-level subresources should be used instead) This field is alpha-level and is only honored by servers that enable the CustomResourceWebhookConversion feature. - subresources:: { - local __subresourcesMixin(subresources) = { subresources+: subresources }, - mixinInstance(subresources):: __subresourcesMixin(subresources), - // Scale denotes the scale subresource for CustomResources - scale:: { - local __scaleMixin(scale) = __subresourcesMixin({ scale+: scale }), - mixinInstance(scale):: __scaleMixin(scale), - // LabelSelectorPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Selector. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. Must be set to work with HPA. If there is no value under the given path in the CustomResource, the status label selector value in the /scale subresource will default to the empty string. - withLabelSelectorPath(labelSelectorPath):: self + __scaleMixin({ labelSelectorPath: labelSelectorPath }), - // SpecReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Spec.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .spec. If there is no value under the given path in the CustomResource, the /scale subresource will return an error on GET. - withSpecReplicasPath(specReplicasPath):: self + __scaleMixin({ specReplicasPath: specReplicasPath }), - // StatusReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. If there is no value under the given path in the CustomResource, the status replica value in the /scale subresource will default to 0. - withStatusReplicasPath(statusReplicasPath):: self + __scaleMixin({ statusReplicasPath: statusReplicasPath }), - }, - scaleType:: hidden.apiextensions.v1beta1.customResourceSubresourceScale, - }, - subresourcesType:: hidden.apiextensions.v1beta1.customResourceSubresources, - }, - }, - // CustomResourceSubresourceScale defines how to serve the scale subresource for CustomResources. - customResourceSubresourceScale:: { - new():: {}, - // LabelSelectorPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Selector. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. Must be set to work with HPA. If there is no value under the given path in the CustomResource, the status label selector value in the /scale subresource will default to the empty string. - withLabelSelectorPath(labelSelectorPath):: self + { labelSelectorPath: labelSelectorPath }, - // SpecReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Spec.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .spec. If there is no value under the given path in the CustomResource, the /scale subresource will return an error on GET. - withSpecReplicasPath(specReplicasPath):: self + { specReplicasPath: specReplicasPath }, - // StatusReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. If there is no value under the given path in the CustomResource, the status replica value in the /scale subresource will default to 0. - withStatusReplicasPath(statusReplicasPath):: self + { statusReplicasPath: statusReplicasPath }, - mixin:: {}, - }, - // CustomResourceSubresourceStatus defines how to serve the status subresource for CustomResources. Status is represented by the `.status` JSON path inside of a CustomResource. When set, * exposes a /status subresource for the custom resource * PUT requests to the /status subresource take a custom resource object, and ignore changes to anything except the status stanza * PUT/POST/PATCH requests to the custom resource ignore changes to the status stanza - customResourceSubresourceStatus:: { - new():: {}, - mixin:: {}, - }, - // CustomResourceSubresources defines the status and scale subresources for CustomResources. - customResourceSubresources:: { - new():: {}, - mixin:: { - // Scale denotes the scale subresource for CustomResources - scale:: { - local __scaleMixin(scale) = { scale+: scale }, - mixinInstance(scale):: __scaleMixin(scale), - // LabelSelectorPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Selector. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. Must be set to work with HPA. If there is no value under the given path in the CustomResource, the status label selector value in the /scale subresource will default to the empty string. - withLabelSelectorPath(labelSelectorPath):: self + __scaleMixin({ labelSelectorPath: labelSelectorPath }), - // SpecReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Spec.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .spec. If there is no value under the given path in the CustomResource, the /scale subresource will return an error on GET. - withSpecReplicasPath(specReplicasPath):: self + __scaleMixin({ specReplicasPath: specReplicasPath }), - // StatusReplicasPath defines the JSON path inside of a CustomResource that corresponds to Scale.Status.Replicas. Only JSON paths without the array notation are allowed. Must be a JSON Path under .status. If there is no value under the given path in the CustomResource, the status replica value in the /scale subresource will default to 0. - withStatusReplicasPath(statusReplicasPath):: self + __scaleMixin({ statusReplicasPath: statusReplicasPath }), - }, - scaleType:: hidden.apiextensions.v1beta1.customResourceSubresourceScale, - }, - }, - // CustomResourceValidation is a list of validation methods for CustomResources. - customResourceValidation:: { - new():: {}, - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3Schema(openApiV3Schema):: self + { openAPIV3Schema: openApiV3Schema }, - // OpenAPIV3Schema is the OpenAPI v3 schema to be validated against. - withOpenApiV3SchemaMixin(openApiV3Schema):: self + { openAPIV3Schema+: openApiV3Schema }, - openAPIV3SchemaType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - mixin:: {}, - }, - // ExternalDocumentation allows referencing an external resource for extended documentation. - externalDocumentation:: { - new():: {}, - withDescription(description):: self + { description: description }, - withUrl(url):: self + { url: url }, - mixin:: {}, - }, - // JSON represents any valid JSON value. These types are supported: bool, int64, float64, string, []interface{}, map[string]interface{} and nil. - json:: { - new():: {}, - mixin:: {}, - }, - // JSONSchemaProps is a JSON-Schema following Specification Draft 4 (http://json-schema.org/). - jsonSchemaProps:: { - new():: {}, - withDollarRef(dollarRef):: self + { "$ref": dollarRef }, - withDollarSchema(dollarSchema):: self + { "$schema": dollarSchema }, - withAllOf(allOf):: self + if std.type(allOf) == 'array' then { allOf: allOf } else { allOf: [allOf] }, - withAllOfMixin(allOf):: self + if std.type(allOf) == 'array' then { allOf+: allOf } else { allOf+: [allOf] }, - allOfType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - withAnyOf(anyOf):: self + if std.type(anyOf) == 'array' then { anyOf: anyOf } else { anyOf: [anyOf] }, - withAnyOfMixin(anyOf):: self + if std.type(anyOf) == 'array' then { anyOf+: anyOf } else { anyOf+: [anyOf] }, - anyOfType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - withDefinitions(definitions):: self + { definitions: definitions }, - withDefinitionsMixin(definitions):: self + { definitions+: definitions }, - withDependencies(dependencies):: self + { dependencies: dependencies }, - withDependenciesMixin(dependencies):: self + { dependencies+: dependencies }, - withDescription(description):: self + { description: description }, - withEnum(enum):: self + if std.type(enum) == 'array' then { enum: enum } else { enum: [enum] }, - withEnumMixin(enum):: self + if std.type(enum) == 'array' then { enum+: enum } else { enum+: [enum] }, - enumType:: hidden.apiextensions.v1beta1.json, - withExclusiveMaximum(exclusiveMaximum):: self + { exclusiveMaximum: exclusiveMaximum }, - withExclusiveMinimum(exclusiveMinimum):: self + { exclusiveMinimum: exclusiveMinimum }, - withFormat(format):: self + { format: format }, - withId(id):: self + { id: id }, - withMaxItems(maxItems):: self + { maxItems: maxItems }, - withMaxLength(maxLength):: self + { maxLength: maxLength }, - withMaxProperties(maxProperties):: self + { maxProperties: maxProperties }, - withMaximum(maximum):: self + { maximum: maximum }, - withMinItems(minItems):: self + { minItems: minItems }, - withMinLength(minLength):: self + { minLength: minLength }, - withMinProperties(minProperties):: self + { minProperties: minProperties }, - withMinimum(minimum):: self + { minimum: minimum }, - withMultipleOf(multipleOf):: self + { multipleOf: multipleOf }, - withNot(not):: self + { not: not }, - withNotMixin(not):: self + { not+: not }, - notType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - withNullable(nullable):: self + { nullable: nullable }, - withOneOf(oneOf):: self + if std.type(oneOf) == 'array' then { oneOf: oneOf } else { oneOf: [oneOf] }, - withOneOfMixin(oneOf):: self + if std.type(oneOf) == 'array' then { oneOf+: oneOf } else { oneOf+: [oneOf] }, - oneOfType:: hidden.apiextensions.v1beta1.jsonSchemaProps, - withPattern(pattern):: self + { pattern: pattern }, - withPatternProperties(patternProperties):: self + { patternProperties: patternProperties }, - withPatternPropertiesMixin(patternProperties):: self + { patternProperties+: patternProperties }, - withProperties(properties):: self + { properties: properties }, - withPropertiesMixin(properties):: self + { properties+: properties }, - withRequired(required):: self + if std.type(required) == 'array' then { required: required } else { required: [required] }, - withRequiredMixin(required):: self + if std.type(required) == 'array' then { required+: required } else { required+: [required] }, - withTitle(title):: self + { title: title }, - withType(type):: self + { type: type }, - withUniqueItems(uniqueItems):: self + { uniqueItems: uniqueItems }, - mixin:: { - additionalItems:: { - local __additionalItemsMixin(additionalItems) = { additionalItems+: additionalItems }, - mixinInstance(additionalItems):: __additionalItemsMixin(additionalItems), - }, - additionalItemsType:: hidden.apiextensions.v1beta1.jsonSchemaPropsOrBool, - additionalProperties:: { - local __additionalPropertiesMixin(additionalProperties) = { additionalProperties+: additionalProperties }, - mixinInstance(additionalProperties):: __additionalPropertiesMixin(additionalProperties), - }, - additionalPropertiesType:: hidden.apiextensions.v1beta1.jsonSchemaPropsOrBool, - default:: { - local __defaultMixin(default) = { default+: default }, - mixinInstance(default):: __defaultMixin(default), - }, - defaultType:: hidden.apiextensions.v1beta1.json, - example:: { - local __exampleMixin(example) = { example+: example }, - mixinInstance(example):: __exampleMixin(example), - }, - exampleType:: hidden.apiextensions.v1beta1.json, - externalDocs:: { - local __externalDocsMixin(externalDocs) = { externalDocs+: externalDocs }, - mixinInstance(externalDocs):: __externalDocsMixin(externalDocs), - withDescription(description):: self + __externalDocsMixin({ description: description }), - withUrl(url):: self + __externalDocsMixin({ url: url }), - }, - externalDocsType:: hidden.apiextensions.v1beta1.externalDocumentation, - items:: { - local __itemsMixin(items) = { items+: items }, - mixinInstance(items):: __itemsMixin(items), - }, - itemsType:: hidden.apiextensions.v1beta1.jsonSchemaPropsOrArray, - }, - }, - // JSONSchemaPropsOrArray represents a value that can either be a JSONSchemaProps or an array of JSONSchemaProps. Mainly here for serialization purposes. - jsonSchemaPropsOrArray:: { - new():: {}, - mixin:: {}, - }, - // JSONSchemaPropsOrBool represents JSONSchemaProps or a boolean value. Defaults to true for the boolean property. - jsonSchemaPropsOrBool:: { - new():: {}, - mixin:: {}, - }, - // JSONSchemaPropsOrStringArray represents a JSONSchemaProps or a string array. - jsonSchemaPropsOrStringArray:: { - new():: {}, - mixin:: {}, - }, - // ServiceReference holds a reference to Service.legacy.k8s.io - serviceReference:: { - new():: {}, - // `name` is the name of the service. Required - withName(name):: self + { name: name }, - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + { namespace: namespace }, - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + { path: path }, - mixin:: {}, - }, - // WebhookClientConfig contains the information to make a TLS connection with the webhook. It has the same field as admissionregistration.v1beta1.WebhookClientConfig. - webhookClientConfig:: { - new():: {}, - // `caBundle` is a PEM encoded CA bundle which will be used to validate the webhook's server certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + { caBundle: caBundle }, - // `url` gives the location of the webhook, in standard URL form (`scheme://host:port/path`). Exactly one of `url` or `service` must be specified. - // - // The `host` should not refer to a service running in the cluster; use the `service` field instead. The host might be resolved via external DNS in some apiservers (e.g., `kube-apiserver` cannot resolve in-cluster DNS as that would be a layering violation). `host` may also be an IP address. - // - // Please note that using `localhost` or `127.0.0.1` as a `host` is risky unless you take great care to run this webhook on all hosts which run an apiserver which might need to make calls to this webhook. Such installs are likely to be non-portable, i.e., not easy to turn up in a new cluster. - // - // The scheme must be "https"; the URL must begin with "https://". - // - // A path is optional, and if present may be any string permissible in a URL. You may use the path to pass an arbitrary string to the webhook, for example, a cluster identifier. - // - // Attempting to use a user or basic auth e.g. "user:password@" is not allowed. Fragments ("#...") and query parameters ("?...") are not allowed, either. - withUrl(url):: self + { url: url }, - mixin:: { - // `service` is a reference to the service for this webhook. Either `service` or `url` must be specified. - // - // If the webhook is running within the cluster, then you should use `service`. - // - // Port 443 will be used if it is open, otherwise it is an error. - service:: { - local __serviceMixin(service) = { service+: service }, - mixinInstance(service):: __serviceMixin(service), - // `name` is the name of the service. Required - withName(name):: self + __serviceMixin({ name: name }), - // `namespace` is the namespace of the service. Required - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - // `path` is an optional URL path which will be sent in any request to this service. - withPath(path):: self + __serviceMixin({ path: path }), - }, - serviceType:: hidden.apiextensions.v1beta1.serviceReference, - }, - }, - }, - }, - apiregistration:: { - v1:: { - local apiVersion = { apiVersion: 'apiregistration/v1' }, - // APIServiceCondition describes the state of an APIService at a particular point - apiServiceCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human-readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // Unique, one-word, CamelCase reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status is the status of the condition. Can be True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type is the type of the condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // APIServiceList is a list of APIService objects. - apiServiceList:: { - new():: {}, - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apiregistration.v1.apiService, - mixin:: {}, - }, - // APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification. - apiServiceSpec:: { - new():: {}, - // CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + { caBundle: caBundle }, - // Group is the API group name this server hosts - withGroup(group):: self + { group: group }, - // GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s - withGroupPriorityMinimum(groupPriorityMinimum):: self + { groupPriorityMinimum: groupPriorityMinimum }, - // InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead. - withInsecureSkipTlsVerify(insecureSkipTlsVerify):: self + { insecureSkipTLSVerify: insecureSkipTlsVerify }, - // Version is the API version this server hosts. For example, "v1" - withVersion(version):: self + { version: version }, - // VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionPriority(versionPriority):: self + { versionPriority: versionPriority }, - mixin:: { - // Service is a reference to the service for this API server. It must communicate on port 443 If the Service is nil, that means the handling for the API groupversion is handled locally on this server. The call will simply delegate to the normal handler chain to be fulfilled. - service:: { - local __serviceMixin(service) = { service+: service }, - mixinInstance(service):: __serviceMixin(service), - // Name is the name of the service - withName(name):: self + __serviceMixin({ name: name }), - // Namespace is the namespace of the service - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - }, - serviceType:: hidden.apiregistration.v1.serviceReference, - }, - }, - // APIServiceStatus contains derived information about an API server - apiServiceStatus:: { - new():: {}, - // Current service state of apiService. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Current service state of apiService. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apiregistration.v1.apiServiceCondition, - mixin:: {}, - }, - // ServiceReference holds a reference to Service.legacy.k8s.io - serviceReference:: { - new():: {}, - // Name is the name of the service - withName(name):: self + { name: name }, - // Namespace is the namespace of the service - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'apiregistration/v1beta1' }, - // APIServiceCondition describes the state of an APIService at a particular point - apiServiceCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human-readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // Unique, one-word, CamelCase reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status is the status of the condition. Can be True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type is the type of the condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // APIServiceList is a list of APIService objects. - apiServiceList:: { - new():: {}, - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apiregistration.v1beta1.apiService, - mixin:: {}, - }, - // APIServiceSpec contains information for locating and communicating with a server. Only https is supported, though you are able to disable certificate verification. - apiServiceSpec:: { - new():: {}, - // CABundle is a PEM encoded CA bundle which will be used to validate an API server's serving certificate. If unspecified, system trust roots on the apiserver are used. - withCaBundle(caBundle):: self + { caBundle: caBundle }, - // Group is the API group name this server hosts - withGroup(group):: self + { group: group }, - // GroupPriorityMininum is the priority this group should have at least. Higher priority means that the group is preferred by clients over lower priority ones. Note that other versions of this group might specify even higher GroupPriorityMininum values such that the whole group gets a higher priority. The primary sort is based on GroupPriorityMinimum, ordered highest number to lowest (20 before 10). The secondary sort is based on the alphabetical comparison of the name of the object. (v1.bar before v1.foo) We'd recommend something like: *.k8s.io (except extensions) at 18000 and PaaSes (OpenShift, Deis) are recommended to be in the 2000s - withGroupPriorityMinimum(groupPriorityMinimum):: self + { groupPriorityMinimum: groupPriorityMinimum }, - // InsecureSkipTLSVerify disables TLS certificate verification when communicating with this server. This is strongly discouraged. You should use the CABundle instead. - withInsecureSkipTlsVerify(insecureSkipTlsVerify):: self + { insecureSkipTLSVerify: insecureSkipTlsVerify }, - // Version is the API version this server hosts. For example, "v1" - withVersion(version):: self + { version: version }, - // VersionPriority controls the ordering of this API version inside of its group. Must be greater than zero. The primary sort is based on VersionPriority, ordered highest to lowest (20 before 10). Since it's inside of a group, the number can be small, probably in the 10s. In case of equal version priorities, the version string will be used to compute the order inside a group. If the version string is "kube-like", it will sort above non "kube-like" version strings, which are ordered lexicographically. "Kube-like" versions start with a "v", then are followed by a number (the major version), then optionally the string "alpha" or "beta" and another number (the minor version). These are sorted first by GA > beta > alpha (where GA is a version with no suffix such as beta or alpha), and then by comparing major version, then minor version. An example sorted list of versions: v10, v2, v1, v11beta2, v10beta3, v3beta1, v12alpha1, v11alpha2, foo1, foo10. - withVersionPriority(versionPriority):: self + { versionPriority: versionPriority }, - mixin:: { - // Service is a reference to the service for this API server. It must communicate on port 443 If the Service is nil, that means the handling for the API groupversion is handled locally on this server. The call will simply delegate to the normal handler chain to be fulfilled. - service:: { - local __serviceMixin(service) = { service+: service }, - mixinInstance(service):: __serviceMixin(service), - // Name is the name of the service - withName(name):: self + __serviceMixin({ name: name }), - // Namespace is the namespace of the service - withNamespace(namespace):: self + __serviceMixin({ namespace: namespace }), - }, - serviceType:: hidden.apiregistration.v1beta1.serviceReference, - }, - }, - // APIServiceStatus contains derived information about an API server - apiServiceStatus:: { - new():: {}, - // Current service state of apiService. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Current service state of apiService. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apiregistration.v1beta1.apiServiceCondition, - mixin:: {}, - }, - // ServiceReference holds a reference to Service.legacy.k8s.io - serviceReference:: { - new():: {}, - // Name is the name of the service - withName(name):: self + { name: name }, - // Namespace is the namespace of the service - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - }, - }, - apps:: { - v1:: { - local apiVersion = { apiVersion: 'apps/v1' }, - // ControllerRevisionList is a resource containing a list of ControllerRevision objects. - controllerRevisionList:: { - new():: {}, - // Items is the list of ControllerRevisions - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of ControllerRevisions - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1.controllerRevision, - mixin:: {}, - }, - // DaemonSetCondition describes the state of a DaemonSet at a certain point. - daemonSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of DaemonSet condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DaemonSetList is a collection of daemon sets. - daemonSetList:: { - new():: {}, - // A list of daemon sets. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // A list of daemon sets. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1.daemonSet, - mixin:: {}, - }, - // DaemonSetSpec is the specification of a daemon set. - daemonSetSpec:: { - new():: {}, - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1.daemonSetUpdateStrategy, - }, - }, - // DaemonSetStatus represents the current status of a daemon set. - daemonSetStatus:: { - new():: {}, - // Count of hash collisions for the DaemonSet. The DaemonSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a DaemonSet's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a DaemonSet's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1.daemonSetCondition, - // The number of nodes that are running at least 1 daemon pod and are supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withCurrentNumberScheduled(currentNumberScheduled):: self + { currentNumberScheduled: currentNumberScheduled }, - // The total number of nodes that should be running the daemon pod (including nodes correctly running the daemon pod). More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withDesiredNumberScheduled(desiredNumberScheduled):: self + { desiredNumberScheduled: desiredNumberScheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberAvailable(numberAvailable):: self + { numberAvailable: numberAvailable }, - // The number of nodes that are running the daemon pod, but are not supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withNumberMisscheduled(numberMisscheduled):: self + { numberMisscheduled: numberMisscheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and ready. - withNumberReady(numberReady):: self + { numberReady: numberReady }, - // The number of nodes that should be running the daemon pod and have none of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberUnavailable(numberUnavailable):: self + { numberUnavailable: numberUnavailable }, - // The most recent generation observed by the daemon set controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The total number of nodes that are running updated daemon pod - withUpdatedNumberScheduled(updatedNumberScheduled):: self + { updatedNumberScheduled: updatedNumberScheduled }, - mixin:: {}, - }, - // DaemonSetUpdateStrategy is a struct used to control the update strategy for a DaemonSet. - daemonSetUpdateStrategy:: { - new():: {}, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDaemonSet, - }, - }, - // DeploymentCondition describes the state of a deployment at a certain point. - deploymentCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // The last time this condition was updated. - withLastUpdateTime(lastUpdateTime):: self + { lastUpdateTime: lastUpdateTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of deployment condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DeploymentList is a list of Deployments. - deploymentList:: { - new(items=''):: self.withItems(items), - // Items is the list of Deployments. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Deployments. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1.deployment, - mixin:: {}, - }, - // DeploymentSpec is the specification of the desired behavior of the Deployment. - deploymentSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Indicates that the deployment is paused. - withPaused(paused):: self + { paused: paused }, - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + { progressDeadlineSeconds: progressDeadlineSeconds }, - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. It must match the pod template's labels. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = { strategy+: strategy }, - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // DeploymentStatus is the most recently observed status of the Deployment. - deploymentStatus:: { - new():: {}, - // Total number of available pods (ready for at least minReadySeconds) targeted by this deployment. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Count of hash collisions for the Deployment. The Deployment controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ReplicaSet. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a deployment's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a deployment's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1.deploymentCondition, - // The generation observed by the deployment controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // Total number of ready pods targeted by this deployment. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Total number of non-terminated pods targeted by this deployment (their labels match the selector). - withReplicas(replicas):: self + { replicas: replicas }, - // Total number of unavailable pods targeted by this deployment. This is the total number of pods that are still required for the deployment to have 100% available capacity. They may either be pods that are running but not yet available or pods that still have not been created. - withUnavailableReplicas(unavailableReplicas):: self + { unavailableReplicas: unavailableReplicas }, - // Total number of non-terminated pods targeted by this deployment that have the desired template spec. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // DeploymentStrategy describes how to replace existing pods with new ones. - deploymentStrategy:: { - new():: {}, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateDeployment, - }, - }, - // ReplicaSetCondition describes the state of a replica set at a certain point. - replicaSetCondition:: { - new():: {}, - // The last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of replica set condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // ReplicaSetList is a collection of ReplicaSets. - replicaSetList:: { - new():: {}, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1.replicaSet, - mixin:: {}, - }, - // ReplicaSetSpec is the specification of a ReplicaSet. - replicaSetSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: { - // Selector is a label query over pods that should match the replica count. Label keys and values that must match in order to be controlled by this replica set. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // ReplicaSetStatus represents the current status of a ReplicaSet. - replicaSetStatus:: { - new():: {}, - // The number of available replicas (ready for at least minReadySeconds) for this replica set. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Represents the latest available observations of a replica set's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a replica set's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1.replicaSetCondition, - // The number of pods that have labels matching the labels of the pod template of the replicaset. - withFullyLabeledReplicas(fullyLabeledReplicas):: self + { fullyLabeledReplicas: fullyLabeledReplicas }, - // ObservedGeneration reflects the generation of the most recently observed ReplicaSet. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The number of ready replicas for this replica set. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // Spec to control the desired behavior of daemon set rolling update. - rollingUpdateDaemonSet:: { - new():: {}, - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // Spec to control the desired behavior of rolling update. - rollingUpdateDeployment:: { - new():: {}, - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + { maxSurge: maxSurge }, - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType. - rollingUpdateStatefulSetStrategy:: { - new():: {}, - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + { partition: partition }, - mixin:: {}, - }, - // StatefulSetCondition describes the state of a statefulset at a certain point. - statefulSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of statefulset condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // StatefulSetList is a collection of StatefulSets. - statefulSetList:: { - new(items=''):: self.withItems(items), - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1.statefulSet, - mixin:: {}, - }, - // A StatefulSetSpec is the specification of a StatefulSet. - statefulSetSpec:: { - new():: {}, - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + { podManagementPolicy: podManagementPolicy }, - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + { serviceName: serviceName }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates: volumeClaimTemplates } else { volumeClaimTemplates: [volumeClaimTemplates] }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates+: volumeClaimTemplates } else { volumeClaimTemplates+: [volumeClaimTemplates] }, - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - mixin:: { - // selector is a label query over pods that should match the replica count. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1.statefulSetUpdateStrategy, - }, - }, - // StatefulSetStatus represents the current state of a StatefulSet. - statefulSetStatus:: { - new():: {}, - // collisionCount is the count of hash collisions for the StatefulSet. The StatefulSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a statefulset's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a statefulset's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1.statefulSetCondition, - // currentReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by currentRevision. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // currentRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [0,currentReplicas). - withCurrentRevision(currentRevision):: self + { currentRevision: currentRevision }, - // observedGeneration is the most recent generation observed for this StatefulSet. It corresponds to the StatefulSet's generation, which is updated on mutation by the API Server. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // readyReplicas is the number of Pods created by the StatefulSet controller that have a Ready Condition. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // replicas is the number of Pods created by the StatefulSet controller. - withReplicas(replicas):: self + { replicas: replicas }, - // updateRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [replicas-updatedReplicas,replicas) - withUpdateRevision(updateRevision):: self + { updateRevision: updateRevision }, - // updatedReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by updateRevision. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy. - statefulSetUpdateStrategy:: { - new():: {}, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1.rollingUpdateStatefulSetStrategy, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'apps/v1beta1' }, - // ControllerRevisionList is a resource containing a list of ControllerRevision objects. - controllerRevisionList:: { - new():: {}, - // Items is the list of ControllerRevisions - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of ControllerRevisions - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta1.controllerRevision, - mixin:: {}, - }, - // DeploymentCondition describes the state of a deployment at a certain point. - deploymentCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // The last time this condition was updated. - withLastUpdateTime(lastUpdateTime):: self + { lastUpdateTime: lastUpdateTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of deployment condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DeploymentList is a list of Deployments. - deploymentList:: { - new(items=''):: self.withItems(items), - // Items is the list of Deployments. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Deployments. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta1.deployment, - mixin:: {}, - }, - // DeploymentSpec is the specification of the desired behavior of the Deployment. - deploymentSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Indicates that the deployment is paused. - withPaused(paused):: self + { paused: paused }, - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + { progressDeadlineSeconds: progressDeadlineSeconds }, - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 2. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // DEPRECATED. The config this deployment is rolling back to. Will be cleared after rollback is done. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = { rollbackTo+: rollbackTo }, - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.apps.v1beta1.rollbackConfig, - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = { strategy+: strategy }, - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1beta1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // DeploymentStatus is the most recently observed status of the Deployment. - deploymentStatus:: { - new():: {}, - // Total number of available pods (ready for at least minReadySeconds) targeted by this deployment. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Count of hash collisions for the Deployment. The Deployment controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ReplicaSet. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a deployment's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a deployment's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta1.deploymentCondition, - // The generation observed by the deployment controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // Total number of ready pods targeted by this deployment. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Total number of non-terminated pods targeted by this deployment (their labels match the selector). - withReplicas(replicas):: self + { replicas: replicas }, - // Total number of unavailable pods targeted by this deployment. This is the total number of pods that are still required for the deployment to have 100% available capacity. They may either be pods that are running but not yet available or pods that still have not been created. - withUnavailableReplicas(unavailableReplicas):: self + { unavailableReplicas: unavailableReplicas }, - // Total number of non-terminated pods targeted by this deployment that have the desired template spec. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // DeploymentStrategy describes how to replace existing pods with new ones. - deploymentStrategy:: { - new():: {}, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateDeployment, - }, - }, - // DEPRECATED. - rollbackConfig:: { - new():: {}, - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + { revision: revision }, - mixin:: {}, - }, - // Spec to control the desired behavior of rolling update. - rollingUpdateDeployment:: { - new():: {}, - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + { maxSurge: maxSurge }, - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType. - rollingUpdateStatefulSetStrategy:: { - new():: {}, - // Partition indicates the ordinal at which the StatefulSet should be partitioned. - withPartition(partition):: self + { partition: partition }, - mixin:: {}, - }, - // ScaleSpec describes the attributes of a scale subresource - scaleSpec:: { - new():: {}, - // desired number of instances for the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // ScaleStatus represents the current status of a scale subresource. - scaleStatus:: { - new():: {}, - // actual number of observed instances of the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelector(selector):: self + { selector: selector }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelectorMixin(selector):: self + { selector+: selector }, - // label selector for pods that should match the replicas count. This is a serializated version of both map-based and more expressive set-based selectors. This is done to avoid introspection in the clients. The string will be in the same format as the query-param syntax. If the target type only supports map-based selectors, both this field and map-based selector field are populated. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withTargetSelector(targetSelector):: self + { targetSelector: targetSelector }, - mixin:: {}, - }, - // StatefulSetCondition describes the state of a statefulset at a certain point. - statefulSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of statefulset condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // StatefulSetList is a collection of StatefulSets. - statefulSetList:: { - new(items=''):: self.withItems(items), - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta1.statefulSet, - mixin:: {}, - }, - // A StatefulSetSpec is the specification of a StatefulSet. - statefulSetSpec:: { - new():: {}, - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + { podManagementPolicy: podManagementPolicy }, - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + { serviceName: serviceName }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates: volumeClaimTemplates } else { volumeClaimTemplates: [volumeClaimTemplates] }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates+: volumeClaimTemplates } else { volumeClaimTemplates+: [volumeClaimTemplates] }, - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - mixin:: { - // selector is a label query over pods that should match the replica count. If empty, defaulted to labels on the pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta1.statefulSetUpdateStrategy, - }, - }, - // StatefulSetStatus represents the current state of a StatefulSet. - statefulSetStatus:: { - new():: {}, - // collisionCount is the count of hash collisions for the StatefulSet. The StatefulSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a statefulset's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a statefulset's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta1.statefulSetCondition, - // currentReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by currentRevision. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // currentRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [0,currentReplicas). - withCurrentRevision(currentRevision):: self + { currentRevision: currentRevision }, - // observedGeneration is the most recent generation observed for this StatefulSet. It corresponds to the StatefulSet's generation, which is updated on mutation by the API Server. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // readyReplicas is the number of Pods created by the StatefulSet controller that have a Ready Condition. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // replicas is the number of Pods created by the StatefulSet controller. - withReplicas(replicas):: self + { replicas: replicas }, - // updateRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [replicas-updatedReplicas,replicas) - withUpdateRevision(updateRevision):: self + { updateRevision: updateRevision }, - // updatedReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by updateRevision. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy. - statefulSetUpdateStrategy:: { - new():: {}, - // Type indicates the type of the StatefulSetUpdateStrategy. - withType(type):: self + { type: type }, - mixin:: { - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta1.rollingUpdateStatefulSetStrategy, - }, - }, - }, - v1beta2:: { - local apiVersion = { apiVersion: 'apps/v1beta2' }, - // ControllerRevisionList is a resource containing a list of ControllerRevision objects. - controllerRevisionList:: { - new():: {}, - // Items is the list of ControllerRevisions - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of ControllerRevisions - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta2.controllerRevision, - mixin:: {}, - }, - // DaemonSetCondition describes the state of a DaemonSet at a certain point. - daemonSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of DaemonSet condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DaemonSetList is a collection of daemon sets. - daemonSetList:: { - new():: {}, - // A list of daemon sets. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // A list of daemon sets. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta2.daemonSet, - mixin:: {}, - }, - // DaemonSetSpec is the specification of a daemon set. - daemonSetSpec:: { - new():: {}, - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta2.daemonSetUpdateStrategy, - }, - }, - // DaemonSetStatus represents the current status of a daemon set. - daemonSetStatus:: { - new():: {}, - // Count of hash collisions for the DaemonSet. The DaemonSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a DaemonSet's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a DaemonSet's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta2.daemonSetCondition, - // The number of nodes that are running at least 1 daemon pod and are supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withCurrentNumberScheduled(currentNumberScheduled):: self + { currentNumberScheduled: currentNumberScheduled }, - // The total number of nodes that should be running the daemon pod (including nodes correctly running the daemon pod). More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withDesiredNumberScheduled(desiredNumberScheduled):: self + { desiredNumberScheduled: desiredNumberScheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberAvailable(numberAvailable):: self + { numberAvailable: numberAvailable }, - // The number of nodes that are running the daemon pod, but are not supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withNumberMisscheduled(numberMisscheduled):: self + { numberMisscheduled: numberMisscheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and ready. - withNumberReady(numberReady):: self + { numberReady: numberReady }, - // The number of nodes that should be running the daemon pod and have none of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberUnavailable(numberUnavailable):: self + { numberUnavailable: numberUnavailable }, - // The most recent generation observed by the daemon set controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The total number of nodes that are running updated daemon pod - withUpdatedNumberScheduled(updatedNumberScheduled):: self + { updatedNumberScheduled: updatedNumberScheduled }, - mixin:: {}, - }, - // DaemonSetUpdateStrategy is a struct used to control the update strategy for a DaemonSet. - daemonSetUpdateStrategy:: { - new():: {}, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDaemonSet, - }, - }, - // DeploymentCondition describes the state of a deployment at a certain point. - deploymentCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // The last time this condition was updated. - withLastUpdateTime(lastUpdateTime):: self + { lastUpdateTime: lastUpdateTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of deployment condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DeploymentList is a list of Deployments. - deploymentList:: { - new(items=''):: self.withItems(items), - // Items is the list of Deployments. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Deployments. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta2.deployment, - mixin:: {}, - }, - // DeploymentSpec is the specification of the desired behavior of the Deployment. - deploymentSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Indicates that the deployment is paused. - withPaused(paused):: self + { paused: paused }, - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. Defaults to 600s. - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + { progressDeadlineSeconds: progressDeadlineSeconds }, - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. It must match the pod template's labels. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = { strategy+: strategy }, - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.apps.v1beta2.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // DeploymentStatus is the most recently observed status of the Deployment. - deploymentStatus:: { - new():: {}, - // Total number of available pods (ready for at least minReadySeconds) targeted by this deployment. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Count of hash collisions for the Deployment. The Deployment controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ReplicaSet. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a deployment's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a deployment's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta2.deploymentCondition, - // The generation observed by the deployment controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // Total number of ready pods targeted by this deployment. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Total number of non-terminated pods targeted by this deployment (their labels match the selector). - withReplicas(replicas):: self + { replicas: replicas }, - // Total number of unavailable pods targeted by this deployment. This is the total number of pods that are still required for the deployment to have 100% available capacity. They may either be pods that are running but not yet available or pods that still have not been created. - withUnavailableReplicas(unavailableReplicas):: self + { unavailableReplicas: unavailableReplicas }, - // Total number of non-terminated pods targeted by this deployment that have the desired template spec. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // DeploymentStrategy describes how to replace existing pods with new ones. - deploymentStrategy:: { - new():: {}, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateDeployment, - }, - }, - // ReplicaSetCondition describes the state of a replica set at a certain point. - replicaSetCondition:: { - new():: {}, - // The last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of replica set condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // ReplicaSetList is a collection of ReplicaSets. - replicaSetList:: { - new():: {}, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta2.replicaSet, - mixin:: {}, - }, - // ReplicaSetSpec is the specification of a ReplicaSet. - replicaSetSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: { - // Selector is a label query over pods that should match the replica count. Label keys and values that must match in order to be controlled by this replica set. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // ReplicaSetStatus represents the current status of a ReplicaSet. - replicaSetStatus:: { - new():: {}, - // The number of available replicas (ready for at least minReadySeconds) for this replica set. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Represents the latest available observations of a replica set's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a replica set's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta2.replicaSetCondition, - // The number of pods that have labels matching the labels of the pod template of the replicaset. - withFullyLabeledReplicas(fullyLabeledReplicas):: self + { fullyLabeledReplicas: fullyLabeledReplicas }, - // ObservedGeneration reflects the generation of the most recently observed ReplicaSet. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The number of ready replicas for this replica set. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // Spec to control the desired behavior of daemon set rolling update. - rollingUpdateDaemonSet:: { - new():: {}, - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // Spec to control the desired behavior of rolling update. - rollingUpdateDeployment:: { - new():: {}, - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + { maxSurge: maxSurge }, - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // RollingUpdateStatefulSetStrategy is used to communicate parameter for RollingUpdateStatefulSetStrategyType. - rollingUpdateStatefulSetStrategy:: { - new():: {}, - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + { partition: partition }, - mixin:: {}, - }, - // ScaleSpec describes the attributes of a scale subresource - scaleSpec:: { - new():: {}, - // desired number of instances for the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // ScaleStatus represents the current status of a scale subresource. - scaleStatus:: { - new():: {}, - // actual number of observed instances of the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelector(selector):: self + { selector: selector }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelectorMixin(selector):: self + { selector+: selector }, - // label selector for pods that should match the replicas count. This is a serializated version of both map-based and more expressive set-based selectors. This is done to avoid introspection in the clients. The string will be in the same format as the query-param syntax. If the target type only supports map-based selectors, both this field and map-based selector field are populated. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withTargetSelector(targetSelector):: self + { targetSelector: targetSelector }, - mixin:: {}, - }, - // StatefulSetCondition describes the state of a statefulset at a certain point. - statefulSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of statefulset condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // StatefulSetList is a collection of StatefulSets. - statefulSetList:: { - new(items=''):: self.withItems(items), - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.apps.v1beta2.statefulSet, - mixin:: {}, - }, - // A StatefulSetSpec is the specification of a StatefulSet. - statefulSetSpec:: { - new():: {}, - // podManagementPolicy controls how pods are created during initial scale up, when replacing pods on nodes, or when scaling down. The default policy is `OrderedReady`, where pods are created in increasing order (pod-0, then pod-1, etc) and the controller will wait until each pod is ready before continuing. When scaling down, the pods are removed in the opposite order. The alternative policy is `Parallel` which will create pods in parallel to match the desired scale without waiting, and on scale down will delete all pods at once. - withPodManagementPolicy(podManagementPolicy):: self + { podManagementPolicy: podManagementPolicy }, - // replicas is the desired number of replicas of the given Template. These are replicas in the sense that they are instantiations of the same Template, but individual replicas also have a consistent identity. If unspecified, defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // revisionHistoryLimit is the maximum number of revisions that will be maintained in the StatefulSet's revision history. The revision history consists of all revisions not represented by a currently applied StatefulSetSpec version. The default value is 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - // serviceName is the name of the service that governs this StatefulSet. This service must exist before the StatefulSet, and is responsible for the network identity of the set. Pods get DNS/hostnames that follow the pattern: pod-specific-string.serviceName.default.svc.cluster.local where "pod-specific-string" is managed by the StatefulSet controller. - withServiceName(serviceName):: self + { serviceName: serviceName }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplates(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates: volumeClaimTemplates } else { volumeClaimTemplates: [volumeClaimTemplates] }, - // volumeClaimTemplates is a list of claims that pods are allowed to reference. The StatefulSet controller is responsible for mapping network identities to claims in a way that maintains the identity of a pod. Every claim in this list must have at least one matching (by name) volumeMount in one container in the template. A claim in this list takes precedence over any volumes in the template, with the same name. - withVolumeClaimTemplatesMixin(volumeClaimTemplates):: self + if std.type(volumeClaimTemplates) == 'array' then { volumeClaimTemplates+: volumeClaimTemplates } else { volumeClaimTemplates+: [volumeClaimTemplates] }, - volumeClaimTemplatesType:: hidden.core.v1.persistentVolumeClaim, - mixin:: { - // selector is a label query over pods that should match the replica count. It must match the pod template's labels. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // template is the object that describes the pod that will be created if insufficient replicas are detected. Each pod stamped out by the StatefulSet will fulfill this Template, but have a unique identity from the rest of the StatefulSet. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // updateStrategy indicates the StatefulSetUpdateStrategy that will be employed to update Pods in the StatefulSet when a revision is made to Template. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateStatefulSetStrategy, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.apps.v1beta2.statefulSetUpdateStrategy, - }, - }, - // StatefulSetStatus represents the current state of a StatefulSet. - statefulSetStatus:: { - new():: {}, - // collisionCount is the count of hash collisions for the StatefulSet. The StatefulSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a statefulset's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a statefulset's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.apps.v1beta2.statefulSetCondition, - // currentReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by currentRevision. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // currentRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [0,currentReplicas). - withCurrentRevision(currentRevision):: self + { currentRevision: currentRevision }, - // observedGeneration is the most recent generation observed for this StatefulSet. It corresponds to the StatefulSet's generation, which is updated on mutation by the API Server. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // readyReplicas is the number of Pods created by the StatefulSet controller that have a Ready Condition. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // replicas is the number of Pods created by the StatefulSet controller. - withReplicas(replicas):: self + { replicas: replicas }, - // updateRevision, if not empty, indicates the version of the StatefulSet used to generate Pods in the sequence [replicas-updatedReplicas,replicas) - withUpdateRevision(updateRevision):: self + { updateRevision: updateRevision }, - // updatedReplicas is the number of Pods created by the StatefulSet controller from the StatefulSet version indicated by updateRevision. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // StatefulSetUpdateStrategy indicates the strategy that the StatefulSet controller will use to perform updates. It includes any additional parameters necessary to perform the update for the indicated strategy. - statefulSetUpdateStrategy:: { - new():: {}, - // Type indicates the type of the StatefulSetUpdateStrategy. Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // RollingUpdate is used to communicate parameters when Type is RollingUpdateStatefulSetStrategyType. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // Partition indicates the ordinal at which the StatefulSet should be partitioned. Default value is 0. - withPartition(partition):: self + __rollingUpdateMixin({ partition: partition }), - }, - rollingUpdateType:: hidden.apps.v1beta2.rollingUpdateStatefulSetStrategy, - }, - }, - }, - }, - authentication:: { - v1:: { - local apiVersion = { apiVersion: 'authentication/v1' }, - // TokenReviewSpec is a description of the token authentication request. - tokenReviewSpec:: { - new():: {}, - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then { audiences: audiences } else { audiences: [audiences] }, - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then { audiences+: audiences } else { audiences+: [audiences] }, - // Token is the opaque bearer token. - withToken(token):: self + { token: token }, - mixin:: {}, - }, - // TokenReviewStatus is the result of the token authentication request. - tokenReviewStatus:: { - new():: {}, - // Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is "true", the token is valid against the audience of the Kubernetes API server. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then { audiences: audiences } else { audiences: [audiences] }, - // Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is "true", the token is valid against the audience of the Kubernetes API server. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then { audiences+: audiences } else { audiences+: [audiences] }, - // Authenticated indicates that the token was associated with a known user. - withAuthenticated(authenticated):: self + { authenticated: authenticated }, - // Error indicates that the token couldn't be checked - withErrorParam(errorParam):: self + { "error": errorParam }, - mixin:: { - // User is the UserInfo associated with the provided token. - user:: { - local __userMixin(user) = { user+: user }, - mixinInstance(user):: __userMixin(user), - // Any additional information provided by the authenticator. - withExtra(extra):: self + __userMixin({ extra: extra }), - // Any additional information provided by the authenticator. - withExtraMixin(extra):: self + __userMixin({ extra+: extra }), - // The names of groups this user is a part of. - withGroups(groups):: self + if std.type(groups) == 'array' then __userMixin({ groups: groups }) else __userMixin({ groups: [groups] }), - // The names of groups this user is a part of. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then __userMixin({ groups+: groups }) else __userMixin({ groups+: [groups] }), - // A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs. - withUid(uid):: self + __userMixin({ uid: uid }), - // The name that uniquely identifies this user among all active users. - withUsername(username):: self + __userMixin({ username: username }), - }, - userType:: hidden.authentication.v1.userInfo, - }, - }, - // UserInfo holds the information about the user needed to implement the user.Info interface. - userInfo:: { - new():: {}, - // Any additional information provided by the authenticator. - withExtra(extra):: self + { extra: extra }, - // Any additional information provided by the authenticator. - withExtraMixin(extra):: self + { extra+: extra }, - // The names of groups this user is a part of. - withGroups(groups):: self + if std.type(groups) == 'array' then { groups: groups } else { groups: [groups] }, - // The names of groups this user is a part of. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then { groups+: groups } else { groups+: [groups] }, - // A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs. - withUid(uid):: self + { uid: uid }, - // The name that uniquely identifies this user among all active users. - withUsername(username):: self + { username: username }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'authentication/v1beta1' }, - // TokenReviewSpec is a description of the token authentication request. - tokenReviewSpec:: { - new():: {}, - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then { audiences: audiences } else { audiences: [audiences] }, - // Audiences is a list of the identifiers that the resource server presented with the token identifies as. Audience-aware token authenticators will verify that the token was intended for at least one of the audiences in this list. If no audiences are provided, the audience will default to the audience of the Kubernetes apiserver. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then { audiences+: audiences } else { audiences+: [audiences] }, - // Token is the opaque bearer token. - withToken(token):: self + { token: token }, - mixin:: {}, - }, - // TokenReviewStatus is the result of the token authentication request. - tokenReviewStatus:: { - new():: {}, - // Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is "true", the token is valid against the audience of the Kubernetes API server. - withAudiences(audiences):: self + if std.type(audiences) == 'array' then { audiences: audiences } else { audiences: [audiences] }, - // Audiences are audience identifiers chosen by the authenticator that are compatible with both the TokenReview and token. An identifier is any identifier in the intersection of the TokenReviewSpec audiences and the token's audiences. A client of the TokenReview API that sets the spec.audiences field should validate that a compatible audience identifier is returned in the status.audiences field to ensure that the TokenReview server is audience aware. If a TokenReview returns an empty status.audience field where status.authenticated is "true", the token is valid against the audience of the Kubernetes API server. - withAudiencesMixin(audiences):: self + if std.type(audiences) == 'array' then { audiences+: audiences } else { audiences+: [audiences] }, - // Authenticated indicates that the token was associated with a known user. - withAuthenticated(authenticated):: self + { authenticated: authenticated }, - // Error indicates that the token couldn't be checked - withErrorParam(errorParam):: self + { "error": errorParam }, - mixin:: { - // User is the UserInfo associated with the provided token. - user:: { - local __userMixin(user) = { user+: user }, - mixinInstance(user):: __userMixin(user), - // Any additional information provided by the authenticator. - withExtra(extra):: self + __userMixin({ extra: extra }), - // Any additional information provided by the authenticator. - withExtraMixin(extra):: self + __userMixin({ extra+: extra }), - // The names of groups this user is a part of. - withGroups(groups):: self + if std.type(groups) == 'array' then __userMixin({ groups: groups }) else __userMixin({ groups: [groups] }), - // The names of groups this user is a part of. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then __userMixin({ groups+: groups }) else __userMixin({ groups+: [groups] }), - // A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs. - withUid(uid):: self + __userMixin({ uid: uid }), - // The name that uniquely identifies this user among all active users. - withUsername(username):: self + __userMixin({ username: username }), - }, - userType:: hidden.authentication.v1beta1.userInfo, - }, - }, - // UserInfo holds the information about the user needed to implement the user.Info interface. - userInfo:: { - new():: {}, - // Any additional information provided by the authenticator. - withExtra(extra):: self + { extra: extra }, - // Any additional information provided by the authenticator. - withExtraMixin(extra):: self + { extra+: extra }, - // The names of groups this user is a part of. - withGroups(groups):: self + if std.type(groups) == 'array' then { groups: groups } else { groups: [groups] }, - // The names of groups this user is a part of. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then { groups+: groups } else { groups+: [groups] }, - // A unique value that identifies this user across time. If this user is deleted and another user by the same name is added, they will have different UIDs. - withUid(uid):: self + { uid: uid }, - // The name that uniquely identifies this user among all active users. - withUsername(username):: self + { username: username }, - mixin:: {}, - }, - }, - }, - authorization:: { - v1:: { - local apiVersion = { apiVersion: 'authorization/v1' }, - // NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface - nonResourceAttributes:: { - new():: {}, - // Path is the URL path of the request - withPath(path):: self + { path: path }, - // Verb is the standard HTTP verb - withVerb(verb):: self + { verb: verb }, - mixin:: {}, - }, - // NonResourceRule holds information that describes a rule for the non-resource - nonResourceRule:: { - new():: {}, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all. - withNonResourceUrls(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs: nonResourceUrls } else { nonResourceURLs: [nonResourceUrls] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all. - withNonResourceUrlsMixin(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs+: nonResourceUrls } else { nonResourceURLs+: [nonResourceUrls] }, - // Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface - resourceAttributes:: { - new():: {}, - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + { group: group }, - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + { name: name }, - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + { namespace: namespace }, - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + { resource: resource }, - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + { subresource: subresource }, - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + { verb: verb }, - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + { version: version }, - mixin:: {}, - }, - // ResourceRule is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - resourceRule:: { - new():: {}, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all. - withApiGroups(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups: apiGroups } else { apiGroups: [apiGroups] }, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all. - withApiGroupsMixin(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups+: apiGroups } else { apiGroups+: [apiGroups] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all. - withResourceNames(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames: resourceNames } else { resourceNames: [resourceNames] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all. - withResourceNamesMixin(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames+: resourceNames } else { resourceNames+: [resourceNames] }, - // Resources is a list of resources this rule applies to. "*" means all in the specified apiGroups. - // "*/foo" represents the subresource 'foo' for all resources in the specified apiGroups. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // Resources is a list of resources this rule applies to. "*" means all in the specified apiGroups. - // "*/foo" represents the subresource 'foo' for all resources in the specified apiGroups. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - // Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set - selfSubjectAccessReviewSpec:: { - new():: {}, - mixin:: { - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = { nonResourceAttributes+: nonResourceAttributes }, - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = { resourceAttributes+: resourceAttributes }, - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1.resourceAttributes, - }, - }, - selfSubjectRulesReviewSpec:: { - new():: {}, - // Namespace to evaluate rules for. Required. - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - // SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set - subjectAccessReviewSpec:: { - new():: {}, - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + { extra: extra }, - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + { extra+: extra }, - // Groups is the groups you're testing for. - withGroups(groups):: self + if std.type(groups) == 'array' then { groups: groups } else { groups: [groups] }, - // Groups is the groups you're testing for. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then { groups+: groups } else { groups+: [groups] }, - // UID information about the requesting user. - withUid(uid):: self + { uid: uid }, - // User is the user you're testing for. If you specify "User" but not "Groups", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + { user: user }, - mixin:: { - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = { nonResourceAttributes+: nonResourceAttributes }, - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = { resourceAttributes+: resourceAttributes }, - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1.resourceAttributes, - }, - }, - // SubjectAccessReviewStatus - subjectAccessReviewStatus:: { - new():: {}, - // Allowed is required. True if the action would be allowed, false otherwise. - withAllowed(allowed):: self + { allowed: allowed }, - // Denied is optional. True if the action would be denied, otherwise false. If both allowed is false and denied is false, then the authorizer has no opinion on whether to authorize the action. Denied may not be true if Allowed is true. - withDenied(denied):: self + { denied: denied }, - // EvaluationError is an indication that some error occurred during the authorization check. It is entirely possible to get an error and be able to continue determine authorization status in spite of it. For instance, RBAC can be missing a role, but enough roles are still present and bound to reason about the request. - withEvaluationError(evaluationError):: self + { evaluationError: evaluationError }, - // Reason is optional. It indicates why a request was allowed or denied. - withReason(reason):: self + { reason: reason }, - mixin:: {}, - }, - // SubjectRulesReviewStatus contains the result of a rules check. This check can be incomplete depending on the set of authorizers the server is configured with and any errors experienced during evaluation. Because authorization rules are additive, if a rule appears in a list it's safe to assume the subject has that permission, even if that list is incomplete. - subjectRulesReviewStatus:: { - new():: {}, - // EvaluationError can appear in combination with Rules. It indicates an error occurred during rule evaluation, such as an authorizer that doesn't support rule evaluation, and that ResourceRules and/or NonResourceRules may be incomplete. - withEvaluationError(evaluationError):: self + { evaluationError: evaluationError }, - // Incomplete is true when the rules returned by this call are incomplete. This is most commonly encountered when an authorizer, such as an external authorizer, doesn't support rules evaluation. - withIncomplete(incomplete):: self + { incomplete: incomplete }, - // NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withNonResourceRules(nonResourceRules):: self + if std.type(nonResourceRules) == 'array' then { nonResourceRules: nonResourceRules } else { nonResourceRules: [nonResourceRules] }, - // NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withNonResourceRulesMixin(nonResourceRules):: self + if std.type(nonResourceRules) == 'array' then { nonResourceRules+: nonResourceRules } else { nonResourceRules+: [nonResourceRules] }, - nonResourceRulesType:: hidden.authorization.v1.nonResourceRule, - // ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withResourceRules(resourceRules):: self + if std.type(resourceRules) == 'array' then { resourceRules: resourceRules } else { resourceRules: [resourceRules] }, - // ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withResourceRulesMixin(resourceRules):: self + if std.type(resourceRules) == 'array' then { resourceRules+: resourceRules } else { resourceRules+: [resourceRules] }, - resourceRulesType:: hidden.authorization.v1.resourceRule, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'authorization/v1beta1' }, - // NonResourceAttributes includes the authorization attributes available for non-resource requests to the Authorizer interface - nonResourceAttributes:: { - new():: {}, - // Path is the URL path of the request - withPath(path):: self + { path: path }, - // Verb is the standard HTTP verb - withVerb(verb):: self + { verb: verb }, - mixin:: {}, - }, - // NonResourceRule holds information that describes a rule for the non-resource - nonResourceRule:: { - new():: {}, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all. - withNonResourceUrls(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs: nonResourceUrls } else { nonResourceURLs: [nonResourceUrls] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path. "*" means all. - withNonResourceUrlsMixin(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs+: nonResourceUrls } else { nonResourceURLs+: [nonResourceUrls] }, - // Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verb is a list of kubernetes non-resource API verbs, like: get, post, put, delete, patch, head, options. "*" means all. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // ResourceAttributes includes the authorization attributes available for resource requests to the Authorizer interface - resourceAttributes:: { - new():: {}, - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + { group: group }, - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + { name: name }, - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + { namespace: namespace }, - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + { resource: resource }, - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + { subresource: subresource }, - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + { verb: verb }, - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + { version: version }, - mixin:: {}, - }, - // ResourceRule is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - resourceRule:: { - new():: {}, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all. - withApiGroups(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups: apiGroups } else { apiGroups: [apiGroups] }, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. "*" means all. - withApiGroupsMixin(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups+: apiGroups } else { apiGroups+: [apiGroups] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all. - withResourceNames(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames: resourceNames } else { resourceNames: [resourceNames] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. "*" means all. - withResourceNamesMixin(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames+: resourceNames } else { resourceNames+: [resourceNames] }, - // Resources is a list of resources this rule applies to. "*" means all in the specified apiGroups. - // "*/foo" represents the subresource 'foo' for all resources in the specified apiGroups. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // Resources is a list of resources this rule applies to. "*" means all in the specified apiGroups. - // "*/foo" represents the subresource 'foo' for all resources in the specified apiGroups. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - // Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verb is a list of kubernetes resource API verbs, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // SelfSubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set - selfSubjectAccessReviewSpec:: { - new():: {}, - mixin:: { - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = { nonResourceAttributes+: nonResourceAttributes }, - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1beta1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = { resourceAttributes+: resourceAttributes }, - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1beta1.resourceAttributes, - }, - }, - selfSubjectRulesReviewSpec:: { - new():: {}, - // Namespace to evaluate rules for. Required. - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - // SubjectAccessReviewSpec is a description of the access request. Exactly one of ResourceAuthorizationAttributes and NonResourceAuthorizationAttributes must be set - subjectAccessReviewSpec:: { - new():: {}, - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtra(extra):: self + { extra: extra }, - // Extra corresponds to the user.Info.GetExtra() method from the authenticator. Since that is input to the authorizer it needs a reflection here. - withExtraMixin(extra):: self + { extra+: extra }, - // Groups is the groups you're testing for. - withGroup(group):: self + if std.type(group) == 'array' then { group: group } else { group: [group] }, - // Groups is the groups you're testing for. - withGroupMixin(group):: self + if std.type(group) == 'array' then { group+: group } else { group+: [group] }, - // UID information about the requesting user. - withUid(uid):: self + { uid: uid }, - // User is the user you're testing for. If you specify "User" but not "Group", then is it interpreted as "What if User were not a member of any groups - withUser(user):: self + { user: user }, - mixin:: { - // NonResourceAttributes describes information for a non-resource access request - nonResourceAttributes:: { - local __nonResourceAttributesMixin(nonResourceAttributes) = { nonResourceAttributes+: nonResourceAttributes }, - mixinInstance(nonResourceAttributes):: __nonResourceAttributesMixin(nonResourceAttributes), - // Path is the URL path of the request - withPath(path):: self + __nonResourceAttributesMixin({ path: path }), - // Verb is the standard HTTP verb - withVerb(verb):: self + __nonResourceAttributesMixin({ verb: verb }), - }, - nonResourceAttributesType:: hidden.authorization.v1beta1.nonResourceAttributes, - // ResourceAuthorizationAttributes describes information for a resource access request - resourceAttributes:: { - local __resourceAttributesMixin(resourceAttributes) = { resourceAttributes+: resourceAttributes }, - mixinInstance(resourceAttributes):: __resourceAttributesMixin(resourceAttributes), - // Group is the API Group of the Resource. "*" means all. - withGroup(group):: self + __resourceAttributesMixin({ group: group }), - // Name is the name of the resource being requested for a "get" or deleted for a "delete". "" (empty) means all. - withName(name):: self + __resourceAttributesMixin({ name: name }), - // Namespace is the namespace of the action being requested. Currently, there is no distinction between no namespace and all namespaces "" (empty) is defaulted for LocalSubjectAccessReviews "" (empty) is empty for cluster-scoped resources "" (empty) means "all" for namespace scoped resources from a SubjectAccessReview or SelfSubjectAccessReview - withNamespace(namespace):: self + __resourceAttributesMixin({ namespace: namespace }), - // Resource is one of the existing resource types. "*" means all. - withResource(resource):: self + __resourceAttributesMixin({ resource: resource }), - // Subresource is one of the existing resource types. "" means none. - withSubresource(subresource):: self + __resourceAttributesMixin({ subresource: subresource }), - // Verb is a kubernetes resource API verb, like: get, list, watch, create, update, delete, proxy. "*" means all. - withVerb(verb):: self + __resourceAttributesMixin({ verb: verb }), - // Version is the API Version of the Resource. "*" means all. - withVersion(version):: self + __resourceAttributesMixin({ version: version }), - }, - resourceAttributesType:: hidden.authorization.v1beta1.resourceAttributes, - }, - }, - // SubjectAccessReviewStatus - subjectAccessReviewStatus:: { - new():: {}, - // Allowed is required. True if the action would be allowed, false otherwise. - withAllowed(allowed):: self + { allowed: allowed }, - // Denied is optional. True if the action would be denied, otherwise false. If both allowed is false and denied is false, then the authorizer has no opinion on whether to authorize the action. Denied may not be true if Allowed is true. - withDenied(denied):: self + { denied: denied }, - // EvaluationError is an indication that some error occurred during the authorization check. It is entirely possible to get an error and be able to continue determine authorization status in spite of it. For instance, RBAC can be missing a role, but enough roles are still present and bound to reason about the request. - withEvaluationError(evaluationError):: self + { evaluationError: evaluationError }, - // Reason is optional. It indicates why a request was allowed or denied. - withReason(reason):: self + { reason: reason }, - mixin:: {}, - }, - // SubjectRulesReviewStatus contains the result of a rules check. This check can be incomplete depending on the set of authorizers the server is configured with and any errors experienced during evaluation. Because authorization rules are additive, if a rule appears in a list it's safe to assume the subject has that permission, even if that list is incomplete. - subjectRulesReviewStatus:: { - new():: {}, - // EvaluationError can appear in combination with Rules. It indicates an error occurred during rule evaluation, such as an authorizer that doesn't support rule evaluation, and that ResourceRules and/or NonResourceRules may be incomplete. - withEvaluationError(evaluationError):: self + { evaluationError: evaluationError }, - // Incomplete is true when the rules returned by this call are incomplete. This is most commonly encountered when an authorizer, such as an external authorizer, doesn't support rules evaluation. - withIncomplete(incomplete):: self + { incomplete: incomplete }, - // NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withNonResourceRules(nonResourceRules):: self + if std.type(nonResourceRules) == 'array' then { nonResourceRules: nonResourceRules } else { nonResourceRules: [nonResourceRules] }, - // NonResourceRules is the list of actions the subject is allowed to perform on non-resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withNonResourceRulesMixin(nonResourceRules):: self + if std.type(nonResourceRules) == 'array' then { nonResourceRules+: nonResourceRules } else { nonResourceRules+: [nonResourceRules] }, - nonResourceRulesType:: hidden.authorization.v1beta1.nonResourceRule, - // ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withResourceRules(resourceRules):: self + if std.type(resourceRules) == 'array' then { resourceRules: resourceRules } else { resourceRules: [resourceRules] }, - // ResourceRules is the list of actions the subject is allowed to perform on resources. The list ordering isn't significant, may contain duplicates, and possibly be incomplete. - withResourceRulesMixin(resourceRules):: self + if std.type(resourceRules) == 'array' then { resourceRules+: resourceRules } else { resourceRules+: [resourceRules] }, - resourceRulesType:: hidden.authorization.v1beta1.resourceRule, - mixin:: {}, - }, - }, - }, - autoscaling:: { - v1:: { - local apiVersion = { apiVersion: 'autoscaling/v1' }, - // CrossVersionObjectReference contains enough information to let you identify the referred resource. - crossVersionObjectReference:: { - new():: {}, - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + { kind: kind }, - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + { name: name }, - mixin:: {}, - }, - // list of horizontal pod autoscaler objects. - horizontalPodAutoscalerList:: { - new(items=''):: self.withItems(items), - // list of horizontal pod autoscaler objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // list of horizontal pod autoscaler objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.autoscaling.v1.horizontalPodAutoscaler, - mixin:: {}, - }, - // specification of a horizontal pod autoscaler. - horizontalPodAutoscalerSpec:: { - new():: {}, - // upper limit for the number of pods that can be set by the autoscaler; cannot be smaller than MinReplicas. - withMaxReplicas(maxReplicas):: self + { maxReplicas: maxReplicas }, - // lower limit for the number of pods that can be set by the autoscaler, default 1. - withMinReplicas(minReplicas):: self + { minReplicas: minReplicas }, - // target average CPU utilization (represented as a percentage of requested CPU) over all the pods; if not specified the default autoscaling policy will be used. - withTargetCpuUtilizationPercentage(targetCpuUtilizationPercentage):: self + { targetCPUUtilizationPercentage: targetCpuUtilizationPercentage }, - mixin:: { - // reference to scaled resource; horizontal pod autoscaler will learn the current resource consumption and will set the desired number of pods by using its Scale subresource. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = { scaleTargetRef+: scaleTargetRef }, - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v1.crossVersionObjectReference, - }, - }, - // current status of a horizontal pod autoscaler - horizontalPodAutoscalerStatus:: { - new():: {}, - // current average CPU utilization over all pods, represented as a percentage of requested CPU, e.g. 70 means that an average pod is using now 70% of its requested CPU. - withCurrentCpuUtilizationPercentage(currentCpuUtilizationPercentage):: self + { currentCPUUtilizationPercentage: currentCpuUtilizationPercentage }, - // current number of replicas of pods managed by this autoscaler. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // desired number of replicas of pods managed by this autoscaler. - withDesiredReplicas(desiredReplicas):: self + { desiredReplicas: desiredReplicas }, - // last time the HorizontalPodAutoscaler scaled the number of pods; used by the autoscaler to control how often the number of pods is changed. - withLastScaleTime(lastScaleTime):: self + { lastScaleTime: lastScaleTime }, - // most recent generation observed by this autoscaler. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - mixin:: {}, - }, - // ScaleSpec describes the attributes of a scale subresource. - scaleSpec:: { - new():: {}, - // desired number of instances for the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // ScaleStatus represents the current status of a scale subresource. - scaleStatus:: { - new():: {}, - // actual number of observed instances of the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - // label query over pods that should match the replicas count. This is same as the label selector but in the string format to avoid introspection by clients. The string will be in the same format as the query-param syntax. More info about label selectors: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelector(selector):: self + { selector: selector }, - mixin:: {}, - }, - }, - v2beta1:: { - local apiVersion = { apiVersion: 'autoscaling/v2beta1' }, - // CrossVersionObjectReference contains enough information to let you identify the referred resource. - crossVersionObjectReference:: { - new():: {}, - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + { kind: kind }, - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + { name: name }, - mixin:: {}, - }, - // ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). Exactly one "target" type should be set. - externalMetricSource:: { - new():: {}, - // metricName is the name of the metric in question. - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // metricSelector is used to identify a specific time series within a given metric. - metricSelector:: { - local __metricSelectorMixin(metricSelector) = { metricSelector+: metricSelector }, - mixinInstance(metricSelector):: __metricSelectorMixin(metricSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions: matchExpressions }) else __metricSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions+: matchExpressions }) else __metricSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __metricSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __metricSelectorMixin({ matchLabels+: matchLabels }), - }, - metricSelectorType:: hidden.meta.v1.labelSelector, - // targetAverageValue is the target per-pod value of global metric (as a quantity). Mutually exclusive with TargetValue. - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = { targetAverageValue+: targetAverageValue }, - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - // targetValue is the target value of the metric (as a quantity). Mutually exclusive with TargetAverageValue. - targetValue:: { - local __targetValueMixin(targetValue) = { targetValue+: targetValue }, - mixinInstance(targetValue):: __targetValueMixin(targetValue), - }, - targetValueType:: hidden.core.resource.quantity, - }, - }, - // ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object. - externalMetricStatus:: { - new():: {}, - // metricName is the name of a metric used for autoscaling in metric system. - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // currentAverageValue is the current value of metric averaged over autoscaled pods. - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = { currentAverageValue+: currentAverageValue }, - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - // currentValue is the current value of the metric (as a quantity) - currentValue:: { - local __currentValueMixin(currentValue) = { currentValue+: currentValue }, - mixinInstance(currentValue):: __currentValueMixin(currentValue), - }, - currentValueType:: hidden.core.resource.quantity, - // metricSelector is used to identify a specific time series within a given metric. - metricSelector:: { - local __metricSelectorMixin(metricSelector) = { metricSelector+: metricSelector }, - mixinInstance(metricSelector):: __metricSelectorMixin(metricSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions: matchExpressions }) else __metricSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions+: matchExpressions }) else __metricSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __metricSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __metricSelectorMixin({ matchLabels+: matchLabels }), - }, - metricSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // HorizontalPodAutoscalerCondition describes the state of a HorizontalPodAutoscaler at a certain point. - horizontalPodAutoscalerCondition:: { - new():: {}, - // lastTransitionTime is the last time the condition transitioned from one status to another - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // message is a human-readable explanation containing details about the transition - withMessage(message):: self + { message: message }, - // reason is the reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // status is the status of the condition (True, False, Unknown) - withStatus(status):: self + { status: status }, - // type describes the current condition - withType(type):: self + { type: type }, - mixin:: {}, - }, - // HorizontalPodAutoscaler is a list of horizontal pod autoscaler objects. - horizontalPodAutoscalerList:: { - new(items=''):: self.withItems(items), - // items is the list of horizontal pod autoscaler objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of horizontal pod autoscaler objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.autoscaling.v2beta1.horizontalPodAutoscaler, - mixin:: {}, - }, - // HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler. - horizontalPodAutoscalerSpec:: { - new():: {}, - // maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas. - withMaxReplicas(maxReplicas):: self + { maxReplicas: maxReplicas }, - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. - withMetrics(metrics):: self + if std.type(metrics) == 'array' then { metrics: metrics } else { metrics: [metrics] }, - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. - withMetricsMixin(metrics):: self + if std.type(metrics) == 'array' then { metrics+: metrics } else { metrics+: [metrics] }, - metricsType:: hidden.autoscaling.v2beta1.metricSpec, - // minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. - withMinReplicas(minReplicas):: self + { minReplicas: minReplicas }, - mixin:: { - // scaleTargetRef points to the target resource to scale, and is used to the pods for which metrics should be collected, as well as to actually change the replica count. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = { scaleTargetRef+: scaleTargetRef }, - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - }, - }, - // HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler. - horizontalPodAutoscalerStatus:: { - new():: {}, - // conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.autoscaling.v2beta1.horizontalPodAutoscalerCondition, - // currentMetrics is the last read state of the metrics used by this autoscaler. - withCurrentMetrics(currentMetrics):: self + if std.type(currentMetrics) == 'array' then { currentMetrics: currentMetrics } else { currentMetrics: [currentMetrics] }, - // currentMetrics is the last read state of the metrics used by this autoscaler. - withCurrentMetricsMixin(currentMetrics):: self + if std.type(currentMetrics) == 'array' then { currentMetrics+: currentMetrics } else { currentMetrics+: [currentMetrics] }, - currentMetricsType:: hidden.autoscaling.v2beta1.metricStatus, - // currentReplicas is current number of replicas of pods managed by this autoscaler, as last seen by the autoscaler. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // desiredReplicas is the desired number of replicas of pods managed by this autoscaler, as last calculated by the autoscaler. - withDesiredReplicas(desiredReplicas):: self + { desiredReplicas: desiredReplicas }, - // lastScaleTime is the last time the HorizontalPodAutoscaler scaled the number of pods, used by the autoscaler to control how often the number of pods is changed. - withLastScaleTime(lastScaleTime):: self + { lastScaleTime: lastScaleTime }, - // observedGeneration is the most recent generation observed by this autoscaler. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - mixin:: {}, - }, - // MetricSpec specifies how to scale based on a single metric (only `type` and one other matching field should be set at once). - metricSpec:: { - new():: {}, - // type is the type of metric source. It should be one of "Object", "Pods" or "Resource", each mapping to a matching field in the object. - withType(type):: self + { type: type }, - mixin:: { - // external refers to a global metric that is not associated with any Kubernetes object. It allows autoscaling based on information coming from components running outside of cluster (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). - external:: { - local __externalMixin(external) = { external+: external }, - mixinInstance(external):: __externalMixin(external), - // metricName is the name of the metric in question. - withMetricName(metricName):: self + __externalMixin({ metricName: metricName }), - // metricSelector is used to identify a specific time series within a given metric. - metricSelector:: { - local __metricSelectorMixin(metricSelector) = __externalMixin({ metricSelector+: metricSelector }), - mixinInstance(metricSelector):: __metricSelectorMixin(metricSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions: matchExpressions }) else __metricSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions+: matchExpressions }) else __metricSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __metricSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __metricSelectorMixin({ matchLabels+: matchLabels }), - }, - metricSelectorType:: hidden.meta.v1.labelSelector, - // targetAverageValue is the target per-pod value of global metric (as a quantity). Mutually exclusive with TargetValue. - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = __externalMixin({ targetAverageValue+: targetAverageValue }), - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - // targetValue is the target value of the metric (as a quantity). Mutually exclusive with TargetAverageValue. - targetValue:: { - local __targetValueMixin(targetValue) = __externalMixin({ targetValue+: targetValue }), - mixinInstance(targetValue):: __targetValueMixin(targetValue), - }, - targetValueType:: hidden.core.resource.quantity, - }, - externalType:: hidden.autoscaling.v2beta1.externalMetricSource, - // object refers to a metric describing a single kubernetes object (for example, hits-per-second on an Ingress object). - object:: { - local __objectMixin(object) = { object+: object }, - mixinInstance(object):: __objectMixin(object), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __objectMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // metricName is the name of the metric in question. - withMetricName(metricName):: self + __objectMixin({ metricName: metricName }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __objectMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // target is the described Kubernetes object. - target:: { - local __targetMixin(target) = __objectMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __targetMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __targetMixin({ name: name }), - }, - targetType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - // targetValue is the target value of the metric (as a quantity). - targetValue:: { - local __targetValueMixin(targetValue) = __objectMixin({ targetValue+: targetValue }), - mixinInstance(targetValue):: __targetValueMixin(targetValue), - }, - targetValueType:: hidden.core.resource.quantity, - }, - objectType:: hidden.autoscaling.v2beta1.objectMetricSource, - // pods refers to a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - pods:: { - local __podsMixin(pods) = { pods+: pods }, - mixinInstance(pods):: __podsMixin(pods), - // metricName is the name of the metric in question - withMetricName(metricName):: self + __podsMixin({ metricName: metricName }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __podsMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // targetAverageValue is the target value of the average of the metric across all relevant pods (as a quantity) - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = __podsMixin({ targetAverageValue+: targetAverageValue }), - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - }, - podsType:: hidden.autoscaling.v2beta1.podsMetricSource, - // resource refers to a resource metric (such as those specified in requests and limits) known to Kubernetes describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resource:: { - local __resourceMixin(resource) = { resource+: resource }, - mixinInstance(resource):: __resourceMixin(resource), - // name is the name of the resource in question. - withName(name):: self + __resourceMixin({ name: name }), - // targetAverageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withTargetAverageUtilization(targetAverageUtilization):: self + __resourceMixin({ targetAverageUtilization: targetAverageUtilization }), - // targetAverageValue is the target value of the average of the resource metric across all relevant pods, as a raw value (instead of as a percentage of the request), similar to the "pods" metric source type. - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = __resourceMixin({ targetAverageValue+: targetAverageValue }), - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - }, - resourceType:: hidden.autoscaling.v2beta1.resourceMetricSource, - }, - }, - // MetricStatus describes the last-read state of a single metric. - metricStatus:: { - new():: {}, - // type is the type of metric source. It will be one of "Object", "Pods" or "Resource", each corresponds to a matching field in the object. - withType(type):: self + { type: type }, - mixin:: { - // external refers to a global metric that is not associated with any Kubernetes object. It allows autoscaling based on information coming from components running outside of cluster (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). - external:: { - local __externalMixin(external) = { external+: external }, - mixinInstance(external):: __externalMixin(external), - // currentAverageValue is the current value of metric averaged over autoscaled pods. - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = __externalMixin({ currentAverageValue+: currentAverageValue }), - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - // currentValue is the current value of the metric (as a quantity) - currentValue:: { - local __currentValueMixin(currentValue) = __externalMixin({ currentValue+: currentValue }), - mixinInstance(currentValue):: __currentValueMixin(currentValue), - }, - currentValueType:: hidden.core.resource.quantity, - // metricName is the name of a metric used for autoscaling in metric system. - withMetricName(metricName):: self + __externalMixin({ metricName: metricName }), - // metricSelector is used to identify a specific time series within a given metric. - metricSelector:: { - local __metricSelectorMixin(metricSelector) = __externalMixin({ metricSelector+: metricSelector }), - mixinInstance(metricSelector):: __metricSelectorMixin(metricSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions: matchExpressions }) else __metricSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __metricSelectorMixin({ matchExpressions+: matchExpressions }) else __metricSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __metricSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __metricSelectorMixin({ matchLabels+: matchLabels }), - }, - metricSelectorType:: hidden.meta.v1.labelSelector, - }, - externalType:: hidden.autoscaling.v2beta1.externalMetricStatus, - // object refers to a metric describing a single kubernetes object (for example, hits-per-second on an Ingress object). - object:: { - local __objectMixin(object) = { object+: object }, - mixinInstance(object):: __objectMixin(object), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __objectMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // currentValue is the current value of the metric (as a quantity). - currentValue:: { - local __currentValueMixin(currentValue) = __objectMixin({ currentValue+: currentValue }), - mixinInstance(currentValue):: __currentValueMixin(currentValue), - }, - currentValueType:: hidden.core.resource.quantity, - // metricName is the name of the metric in question. - withMetricName(metricName):: self + __objectMixin({ metricName: metricName }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set in the ObjectMetricSource, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __objectMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // target is the described Kubernetes object. - target:: { - local __targetMixin(target) = __objectMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __targetMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __targetMixin({ name: name }), - }, - targetType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - }, - objectType:: hidden.autoscaling.v2beta1.objectMetricStatus, - // pods refers to a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - pods:: { - local __podsMixin(pods) = { pods+: pods }, - mixinInstance(pods):: __podsMixin(pods), - // currentAverageValue is the current value of the average of the metric across all relevant pods (as a quantity) - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = __podsMixin({ currentAverageValue+: currentAverageValue }), - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - // metricName is the name of the metric in question - withMetricName(metricName):: self + __podsMixin({ metricName: metricName }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set in the PodsMetricSource, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __podsMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - podsType:: hidden.autoscaling.v2beta1.podsMetricStatus, - // resource refers to a resource metric (such as those specified in requests and limits) known to Kubernetes describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resource:: { - local __resourceMixin(resource) = { resource+: resource }, - mixinInstance(resource):: __resourceMixin(resource), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. It will only be present if `targetAverageValue` was set in the corresponding metric specification. - withCurrentAverageUtilization(currentAverageUtilization):: self + __resourceMixin({ currentAverageUtilization: currentAverageUtilization }), - // currentAverageValue is the current value of the average of the resource metric across all relevant pods, as a raw value (instead of as a percentage of the request), similar to the "pods" metric source type. It will always be set, regardless of the corresponding metric specification. - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = __resourceMixin({ currentAverageValue+: currentAverageValue }), - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - // name is the name of the resource in question. - withName(name):: self + __resourceMixin({ name: name }), - }, - resourceType:: hidden.autoscaling.v2beta1.resourceMetricStatus, - }, - }, - // ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object). - objectMetricSource:: { - new():: {}, - // metricName is the name of the metric in question. - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = { averageValue+: averageValue }, - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // target is the described Kubernetes object. - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __targetMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __targetMixin({ name: name }), - }, - targetType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - // targetValue is the target value of the metric (as a quantity). - targetValue:: { - local __targetValueMixin(targetValue) = { targetValue+: targetValue }, - mixinInstance(targetValue):: __targetValueMixin(targetValue), - }, - targetValueType:: hidden.core.resource.quantity, - }, - }, - // ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object). - objectMetricStatus:: { - new():: {}, - // metricName is the name of the metric in question. - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = { averageValue+: averageValue }, - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // currentValue is the current value of the metric (as a quantity). - currentValue:: { - local __currentValueMixin(currentValue) = { currentValue+: currentValue }, - mixinInstance(currentValue):: __currentValueMixin(currentValue), - }, - currentValueType:: hidden.core.resource.quantity, - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set in the ObjectMetricSource, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // target is the described Kubernetes object. - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __targetMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __targetMixin({ name: name }), - }, - targetType:: hidden.autoscaling.v2beta1.crossVersionObjectReference, - }, - }, - // PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - podsMetricSource:: { - new():: {}, - // metricName is the name of the metric in question - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // targetAverageValue is the target value of the average of the metric across all relevant pods (as a quantity) - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = { targetAverageValue+: targetAverageValue }, - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - }, - }, - // PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second). - podsMetricStatus:: { - new():: {}, - // metricName is the name of the metric in question - withMetricName(metricName):: self + { metricName: metricName }, - mixin:: { - // currentAverageValue is the current value of the average of the metric across all relevant pods (as a quantity) - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = { currentAverageValue+: currentAverageValue }, - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set in the PodsMetricSource, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - }, - // ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set. - resourceMetricSource:: { - new():: {}, - // name is the name of the resource in question. - withName(name):: self + { name: name }, - // targetAverageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withTargetAverageUtilization(targetAverageUtilization):: self + { targetAverageUtilization: targetAverageUtilization }, - mixin:: { - // targetAverageValue is the target value of the average of the resource metric across all relevant pods, as a raw value (instead of as a percentage of the request), similar to the "pods" metric source type. - targetAverageValue:: { - local __targetAverageValueMixin(targetAverageValue) = { targetAverageValue+: targetAverageValue }, - mixinInstance(targetAverageValue):: __targetAverageValueMixin(targetAverageValue), - }, - targetAverageValueType:: hidden.core.resource.quantity, - }, - }, - // ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resourceMetricStatus:: { - new():: {}, - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. It will only be present if `targetAverageValue` was set in the corresponding metric specification. - withCurrentAverageUtilization(currentAverageUtilization):: self + { currentAverageUtilization: currentAverageUtilization }, - // name is the name of the resource in question. - withName(name):: self + { name: name }, - mixin:: { - // currentAverageValue is the current value of the average of the resource metric across all relevant pods, as a raw value (instead of as a percentage of the request), similar to the "pods" metric source type. It will always be set, regardless of the corresponding metric specification. - currentAverageValue:: { - local __currentAverageValueMixin(currentAverageValue) = { currentAverageValue+: currentAverageValue }, - mixinInstance(currentAverageValue):: __currentAverageValueMixin(currentAverageValue), - }, - currentAverageValueType:: hidden.core.resource.quantity, - }, - }, - }, - v2beta2:: { - local apiVersion = { apiVersion: 'autoscaling/v2beta2' }, - // CrossVersionObjectReference contains enough information to let you identify the referred resource. - crossVersionObjectReference:: { - new():: {}, - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + { kind: kind }, - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + { name: name }, - mixin:: {}, - }, - // ExternalMetricSource indicates how to scale on a metric not associated with any Kubernetes object (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). - externalMetricSource:: { - new():: {}, - mixin:: { - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - }, - // ExternalMetricStatus indicates the current value of a global metric not associated with any Kubernetes object. - externalMetricStatus:: { - new():: {}, - mixin:: { - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = { current+: current }, - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - }, - // HorizontalPodAutoscalerCondition describes the state of a HorizontalPodAutoscaler at a certain point. - horizontalPodAutoscalerCondition:: { - new():: {}, - // lastTransitionTime is the last time the condition transitioned from one status to another - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // message is a human-readable explanation containing details about the transition - withMessage(message):: self + { message: message }, - // reason is the reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // status is the status of the condition (True, False, Unknown) - withStatus(status):: self + { status: status }, - // type describes the current condition - withType(type):: self + { type: type }, - mixin:: {}, - }, - // HorizontalPodAutoscalerList is a list of horizontal pod autoscaler objects. - horizontalPodAutoscalerList:: { - new(items=''):: self.withItems(items), - // items is the list of horizontal pod autoscaler objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of horizontal pod autoscaler objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.autoscaling.v2beta2.horizontalPodAutoscaler, - mixin:: {}, - }, - // HorizontalPodAutoscalerSpec describes the desired functionality of the HorizontalPodAutoscaler. - horizontalPodAutoscalerSpec:: { - new():: {}, - // maxReplicas is the upper limit for the number of replicas to which the autoscaler can scale up. It cannot be less that minReplicas. - withMaxReplicas(maxReplicas):: self + { maxReplicas: maxReplicas }, - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization. - withMetrics(metrics):: self + if std.type(metrics) == 'array' then { metrics: metrics } else { metrics: [metrics] }, - // metrics contains the specifications for which to use to calculate the desired replica count (the maximum replica count across all metrics will be used). The desired replica count is calculated multiplying the ratio between the target value and the current value by the current number of pods. Ergo, metrics used must decrease as the pod count is increased, and vice-versa. See the individual metric source types for more information about how each type of metric must respond. If not set, the default metric will be set to 80% average CPU utilization. - withMetricsMixin(metrics):: self + if std.type(metrics) == 'array' then { metrics+: metrics } else { metrics+: [metrics] }, - metricsType:: hidden.autoscaling.v2beta2.metricSpec, - // minReplicas is the lower limit for the number of replicas to which the autoscaler can scale down. It defaults to 1 pod. - withMinReplicas(minReplicas):: self + { minReplicas: minReplicas }, - mixin:: { - // scaleTargetRef points to the target resource to scale, and is used to the pods for which metrics should be collected, as well as to actually change the replica count. - scaleTargetRef:: { - local __scaleTargetRefMixin(scaleTargetRef) = { scaleTargetRef+: scaleTargetRef }, - mixinInstance(scaleTargetRef):: __scaleTargetRefMixin(scaleTargetRef), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __scaleTargetRefMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __scaleTargetRefMixin({ name: name }), - }, - scaleTargetRefType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - }, - }, - // HorizontalPodAutoscalerStatus describes the current status of a horizontal pod autoscaler. - horizontalPodAutoscalerStatus:: { - new():: {}, - // conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // conditions is the set of conditions required for this autoscaler to scale its target, and indicates whether or not those conditions are met. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.autoscaling.v2beta2.horizontalPodAutoscalerCondition, - // currentMetrics is the last read state of the metrics used by this autoscaler. - withCurrentMetrics(currentMetrics):: self + if std.type(currentMetrics) == 'array' then { currentMetrics: currentMetrics } else { currentMetrics: [currentMetrics] }, - // currentMetrics is the last read state of the metrics used by this autoscaler. - withCurrentMetricsMixin(currentMetrics):: self + if std.type(currentMetrics) == 'array' then { currentMetrics+: currentMetrics } else { currentMetrics+: [currentMetrics] }, - currentMetricsType:: hidden.autoscaling.v2beta2.metricStatus, - // currentReplicas is current number of replicas of pods managed by this autoscaler, as last seen by the autoscaler. - withCurrentReplicas(currentReplicas):: self + { currentReplicas: currentReplicas }, - // desiredReplicas is the desired number of replicas of pods managed by this autoscaler, as last calculated by the autoscaler. - withDesiredReplicas(desiredReplicas):: self + { desiredReplicas: desiredReplicas }, - // lastScaleTime is the last time the HorizontalPodAutoscaler scaled the number of pods, used by the autoscaler to control how often the number of pods is changed. - withLastScaleTime(lastScaleTime):: self + { lastScaleTime: lastScaleTime }, - // observedGeneration is the most recent generation observed by this autoscaler. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - mixin:: {}, - }, - // MetricIdentifier defines the name and optionally selector for a metric - metricIdentifier:: { - new():: {}, - // name is the name of the given metric - withName(name):: self + { name: name }, - mixin:: { - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - }, - // MetricSpec specifies how to scale based on a single metric (only `type` and one other matching field should be set at once). - metricSpec:: { - new():: {}, - // type is the type of metric source. It should be one of "Object", "Pods" or "Resource", each mapping to a matching field in the object. - withType(type):: self + { type: type }, - mixin:: { - // external refers to a global metric that is not associated with any Kubernetes object. It allows autoscaling based on information coming from components running outside of cluster (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). - external:: { - local __externalMixin(external) = { external+: external }, - mixinInstance(external):: __externalMixin(external), - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __externalMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = __externalMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - externalType:: hidden.autoscaling.v2beta2.externalMetricSource, - // object refers to a metric describing a single kubernetes object (for example, hits-per-second on an Ingress object). - object:: { - local __objectMixin(object) = { object+: object }, - mixinInstance(object):: __objectMixin(object), - describedObject:: { - local __describedObjectMixin(describedObject) = __objectMixin({ describedObject+: describedObject }), - mixinInstance(describedObject):: __describedObjectMixin(describedObject), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __describedObjectMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __describedObjectMixin({ name: name }), - }, - describedObjectType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __objectMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = __objectMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - objectType:: hidden.autoscaling.v2beta2.objectMetricSource, - // pods refers to a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - pods:: { - local __podsMixin(pods) = { pods+: pods }, - mixinInstance(pods):: __podsMixin(pods), - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __podsMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = __podsMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - podsType:: hidden.autoscaling.v2beta2.podsMetricSource, - // resource refers to a resource metric (such as those specified in requests and limits) known to Kubernetes describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resource:: { - local __resourceMixin(resource) = { resource+: resource }, - mixinInstance(resource):: __resourceMixin(resource), - // name is the name of the resource in question. - withName(name):: self + __resourceMixin({ name: name }), - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = __resourceMixin({ target+: target }), - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - resourceType:: hidden.autoscaling.v2beta2.resourceMetricSource, - }, - }, - // MetricStatus describes the last-read state of a single metric. - metricStatus:: { - new():: {}, - // type is the type of metric source. It will be one of "Object", "Pods" or "Resource", each corresponds to a matching field in the object. - withType(type):: self + { type: type }, - mixin:: { - // external refers to a global metric that is not associated with any Kubernetes object. It allows autoscaling based on information coming from components running outside of cluster (for example length of queue in cloud messaging service, or QPS from loadbalancer running outside of cluster). - external:: { - local __externalMixin(external) = { external+: external }, - mixinInstance(external):: __externalMixin(external), - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = __externalMixin({ current+: current }), - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __externalMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - externalType:: hidden.autoscaling.v2beta2.externalMetricStatus, - // object refers to a metric describing a single kubernetes object (for example, hits-per-second on an Ingress object). - object:: { - local __objectMixin(object) = { object+: object }, - mixinInstance(object):: __objectMixin(object), - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = __objectMixin({ current+: current }), - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - describedObject:: { - local __describedObjectMixin(describedObject) = __objectMixin({ describedObject+: describedObject }), - mixinInstance(describedObject):: __describedObjectMixin(describedObject), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __describedObjectMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __describedObjectMixin({ name: name }), - }, - describedObjectType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __objectMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - objectType:: hidden.autoscaling.v2beta2.objectMetricStatus, - // pods refers to a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - pods:: { - local __podsMixin(pods) = { pods+: pods }, - mixinInstance(pods):: __podsMixin(pods), - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = __podsMixin({ current+: current }), - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = __podsMixin({ metric+: metric }), - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - podsType:: hidden.autoscaling.v2beta2.podsMetricStatus, - // resource refers to a resource metric (such as those specified in requests and limits) known to Kubernetes describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resource:: { - local __resourceMixin(resource) = { resource+: resource }, - mixinInstance(resource):: __resourceMixin(resource), - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = __resourceMixin({ current+: current }), - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - // Name is the name of the resource in question. - withName(name):: self + __resourceMixin({ name: name }), - }, - resourceType:: hidden.autoscaling.v2beta2.resourceMetricStatus, - }, - }, - // MetricTarget defines the target value, average value, or average utilization of a specific metric - metricTarget:: { - new():: {}, - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + { averageUtilization: averageUtilization }, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + { type: type }, - mixin:: { - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = { averageValue+: averageValue }, - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = { value+: value }, - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - }, - // MetricValueStatus holds the current value for a metric - metricValueStatus:: { - new():: {}, - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + { averageUtilization: averageUtilization }, - mixin:: { - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = { averageValue+: averageValue }, - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = { value+: value }, - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - }, - // ObjectMetricSource indicates how to scale on a metric describing a kubernetes object (for example, hits-per-second on an Ingress object). - objectMetricSource:: { - new():: {}, - mixin:: { - describedObject:: { - local __describedObjectMixin(describedObject) = { describedObject+: describedObject }, - mixinInstance(describedObject):: __describedObjectMixin(describedObject), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __describedObjectMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __describedObjectMixin({ name: name }), - }, - describedObjectType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - }, - // ObjectMetricStatus indicates the current value of a metric describing a kubernetes object (for example, hits-per-second on an Ingress object). - objectMetricStatus:: { - new():: {}, - mixin:: { - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = { current+: current }, - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - describedObject:: { - local __describedObjectMixin(describedObject) = { describedObject+: describedObject }, - mixinInstance(describedObject):: __describedObjectMixin(describedObject), - // Kind of the referent; More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds" - withKind(kind):: self + __describedObjectMixin({ kind: kind }), - // Name of the referent; More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __describedObjectMixin({ name: name }), - }, - describedObjectType:: hidden.autoscaling.v2beta2.crossVersionObjectReference, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - }, - // PodsMetricSource indicates how to scale on a metric describing each pod in the current scale target (for example, transactions-processed-per-second). The values will be averaged together before being compared to the target value. - podsMetricSource:: { - new():: {}, - mixin:: { - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - }, - // PodsMetricStatus indicates the current value of a metric describing each pod in the current scale target (for example, transactions-processed-per-second). - podsMetricStatus:: { - new():: {}, - mixin:: { - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = { current+: current }, - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - // metric identifies the target metric by name and selector - metric:: { - local __metricMixin(metric) = { metric+: metric }, - mixinInstance(metric):: __metricMixin(metric), - // name is the name of the given metric - withName(name):: self + __metricMixin({ name: name }), - // selector is the string-encoded form of a standard kubernetes label selector for the given metric When set, it is passed as an additional parameter to the metrics server for more specific metrics scoping. When unset, just the metricName will be used to gather metrics. - selector:: { - local __selectorMixin(selector) = __metricMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - metricType:: hidden.autoscaling.v2beta2.metricIdentifier, - }, - }, - // ResourceMetricSource indicates how to scale on a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). The values will be averaged together before being compared to the target. Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. Only one "target" type should be set. - resourceMetricSource:: { - new():: {}, - // name is the name of the resource in question. - withName(name):: self + { name: name }, - mixin:: { - // target specifies the target value for the given metric - target:: { - local __targetMixin(target) = { target+: target }, - mixinInstance(target):: __targetMixin(target), - // averageUtilization is the target value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. Currently only valid for Resource metric source type - withAverageUtilization(averageUtilization):: self + __targetMixin({ averageUtilization: averageUtilization }), - // averageValue is the target value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __targetMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // type represents whether the metric type is Utilization, Value, or AverageValue - withType(type):: self + __targetMixin({ type: type }), - // value is the target value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __targetMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - targetType:: hidden.autoscaling.v2beta2.metricTarget, - }, - }, - // ResourceMetricStatus indicates the current value of a resource metric known to Kubernetes, as specified in requests and limits, describing each pod in the current scale target (e.g. CPU or memory). Such metrics are built in to Kubernetes, and have special scaling options on top of those available to normal per-pod metrics using the "pods" source. - resourceMetricStatus:: { - new():: {}, - // Name is the name of the resource in question. - withName(name):: self + { name: name }, - mixin:: { - // current contains the current value for the given metric - current:: { - local __currentMixin(current) = { current+: current }, - mixinInstance(current):: __currentMixin(current), - // currentAverageUtilization is the current value of the average of the resource metric across all relevant pods, represented as a percentage of the requested value of the resource for the pods. - withAverageUtilization(averageUtilization):: self + __currentMixin({ averageUtilization: averageUtilization }), - // averageValue is the current value of the average of the metric across all relevant pods (as a quantity) - averageValue:: { - local __averageValueMixin(averageValue) = __currentMixin({ averageValue+: averageValue }), - mixinInstance(averageValue):: __averageValueMixin(averageValue), - }, - averageValueType:: hidden.core.resource.quantity, - // value is the current value of the metric (as a quantity). - value:: { - local __valueMixin(value) = __currentMixin({ value+: value }), - mixinInstance(value):: __valueMixin(value), - }, - valueType:: hidden.core.resource.quantity, - }, - currentType:: hidden.autoscaling.v2beta2.metricValueStatus, - }, - }, - }, - }, - batch:: { - v1:: { - local apiVersion = { apiVersion: 'batch/v1' }, - // JobCondition describes current state of a job. - jobCondition:: { - new():: {}, - // Last time the condition was checked. - withLastProbeTime(lastProbeTime):: self + { lastProbeTime: lastProbeTime }, - // Last time the condition transit from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // (brief) reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of job condition, Complete or Failed. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // JobList is a collection of jobs. - jobList:: { - new(items=''):: self.withItems(items), - // items is the list of Jobs. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of Jobs. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.batch.v1.job, - mixin:: {}, - }, - // JobSpec describes how the job execution will look like. - jobSpec:: { - new():: {}, - // Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + { activeDeadlineSeconds: activeDeadlineSeconds }, - // Specifies the number of retries before marking this job failed. Defaults to 6 - withBackoffLimit(backoffLimit):: self + { backoffLimit: backoffLimit }, - // Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withCompletions(completions):: self + { completions: completions }, - // manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector - withManualSelector(manualSelector):: self + { manualSelector: manualSelector }, - // Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withParallelism(parallelism):: self + { parallelism: parallelism }, - // ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature. - withTtlSecondsAfterFinished(ttlSecondsAfterFinished):: self + { ttlSecondsAfterFinished: ttlSecondsAfterFinished }, - mixin:: { - // A label query over pods that should match the pod count. Normally, the system sets this field for you. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Describes the pod that will be created when executing a job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // JobStatus represents the current state of a Job. - jobStatus:: { - new():: {}, - // The number of actively running pods. - withActive(active):: self + { active: active }, - // Represents time when the job was completed. It is not guaranteed to be set in happens-before order across separate operations. It is represented in RFC3339 form and is in UTC. - withCompletionTime(completionTime):: self + { completionTime: completionTime }, - // The latest available observations of an object's current state. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // The latest available observations of an object's current state. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.batch.v1.jobCondition, - // The number of pods which reached phase Failed. - withFailed(failed):: self + { failed: failed }, - // Represents time when the job was acknowledged by the job controller. It is not guaranteed to be set in happens-before order across separate operations. It is represented in RFC3339 form and is in UTC. - withStartTime(startTime):: self + { startTime: startTime }, - // The number of pods which reached phase Succeeded. - withSucceeded(succeeded):: self + { succeeded: succeeded }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'batch/v1beta1' }, - // CronJobList is a collection of cron jobs. - cronJobList:: { - new(items=''):: self.withItems(items), - // items is the list of CronJobs. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of CronJobs. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.batch.v1beta1.cronJob, - mixin:: {}, - }, - // CronJobSpec describes how the job execution will look like and when it will actually run. - cronJobSpec:: { - new():: {}, - // Specifies how to treat concurrent executions of a Job. Valid values are: - "Allow" (default): allows CronJobs to run concurrently; - "Forbid": forbids concurrent runs, skipping next run if previous run hasn't finished yet; - "Replace": cancels currently running job and replaces it with a new one - withConcurrencyPolicy(concurrencyPolicy):: self + { concurrencyPolicy: concurrencyPolicy }, - // The number of failed finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withFailedJobsHistoryLimit(failedJobsHistoryLimit):: self + { failedJobsHistoryLimit: failedJobsHistoryLimit }, - // The schedule in Cron format, see https://en.wikipedia.org/wiki/Cron. - withSchedule(schedule):: self + { schedule: schedule }, - // Optional deadline in seconds for starting the job if it misses scheduled time for any reason. Missed jobs executions will be counted as failed ones. - withStartingDeadlineSeconds(startingDeadlineSeconds):: self + { startingDeadlineSeconds: startingDeadlineSeconds }, - // The number of successful finished jobs to retain. This is a pointer to distinguish between explicit zero and not specified. Defaults to 3. - withSuccessfulJobsHistoryLimit(successfulJobsHistoryLimit):: self + { successfulJobsHistoryLimit: successfulJobsHistoryLimit }, - // This flag tells the controller to suspend subsequent executions, it does not apply to already started executions. Defaults to false. - withSuspend(suspend):: self + { suspend: suspend }, - mixin:: { - // Specifies the job that will be created when executing a CronJob. - jobTemplate:: { - local __jobTemplateMixin(jobTemplate) = { jobTemplate+: jobTemplate }, - mixinInstance(jobTemplate):: __jobTemplateMixin(jobTemplate), - // Standard object's metadata of the jobs created from this template. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __jobTemplateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the job. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __jobTemplateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // Specifies the number of retries before marking this job failed. Defaults to 6 - withBackoffLimit(backoffLimit):: self + __specMixin({ backoffLimit: backoffLimit }), - // Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withCompletions(completions):: self + __specMixin({ completions: completions }), - // manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector - withManualSelector(manualSelector):: self + __specMixin({ manualSelector: manualSelector }), - // Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withParallelism(parallelism):: self + __specMixin({ parallelism: parallelism }), - // A label query over pods that should match the pod count. Normally, the system sets this field for you. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Describes the pod that will be created when executing a job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature. - withTtlSecondsAfterFinished(ttlSecondsAfterFinished):: self + __specMixin({ ttlSecondsAfterFinished: ttlSecondsAfterFinished }), - }, - specType:: hidden.batch.v1.jobSpec, - }, - jobTemplateType:: hidden.batch.v1beta1.jobTemplateSpec, - }, - }, - // CronJobStatus represents the current state of a cron job. - cronJobStatus:: { - new():: {}, - // A list of pointers to currently running jobs. - withActive(active):: self + if std.type(active) == 'array' then { active: active } else { active: [active] }, - // A list of pointers to currently running jobs. - withActiveMixin(active):: self + if std.type(active) == 'array' then { active+: active } else { active+: [active] }, - activeType:: hidden.core.v1.objectReference, - // Information when was the last time the job was successfully scheduled. - withLastScheduleTime(lastScheduleTime):: self + { lastScheduleTime: lastScheduleTime }, - mixin:: {}, - }, - // JobTemplateSpec describes the data a Job should have when created from a template - jobTemplateSpec:: { - new():: {}, - mixin:: { - // Standard object's metadata of the jobs created from this template. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the job. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Specifies the duration in seconds relative to the startTime that the job may be active before the system tries to terminate it; value must be positive integer - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // Specifies the number of retries before marking this job failed. Defaults to 6 - withBackoffLimit(backoffLimit):: self + __specMixin({ backoffLimit: backoffLimit }), - // Specifies the desired number of successfully finished pods the job should be run with. Setting to nil means that the success of any pod signals the success of all pods, and allows parallelism to have any positive value. Setting to 1 means that parallelism is limited to 1 and the success of that pod signals the success of the job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withCompletions(completions):: self + __specMixin({ completions: completions }), - // manualSelector controls generation of pod labels and pod selectors. Leave `manualSelector` unset unless you are certain what you are doing. When false or unset, the system pick labels unique to this job and appends those labels to the pod template. When true, the user is responsible for picking unique labels and specifying the selector. Failure to pick a unique label may cause this and other jobs to not function correctly. However, You may see `manualSelector=true` in jobs that were created with the old `extensions/v1beta1` API. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/#specifying-your-own-pod-selector - withManualSelector(manualSelector):: self + __specMixin({ manualSelector: manualSelector }), - // Specifies the maximum desired number of pods the job should run at any given time. The actual number of pods running in steady state will be less than this number when ((.spec.completions - .status.successful) < .spec.parallelism), i.e. when the work left to do is less than max parallelism. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - withParallelism(parallelism):: self + __specMixin({ parallelism: parallelism }), - // A label query over pods that should match the pod count. Normally, the system sets this field for you. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = __specMixin({ selector+: selector }), - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Describes the pod that will be created when executing a job. More info: https://kubernetes.io/docs/concepts/workloads/controllers/jobs-run-to-completion/ - template:: { - local __templateMixin(template) = __specMixin({ template+: template }), - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // ttlSecondsAfterFinished limits the lifetime of a Job that has finished execution (either Complete or Failed). If this field is set, ttlSecondsAfterFinished after the Job finishes, it is eligible to be automatically deleted. When the Job is being deleted, its lifecycle guarantees (e.g. finalizers) will be honored. If this field is unset, the Job won't be automatically deleted. If this field is set to zero, the Job becomes eligible to be deleted immediately after it finishes. This field is alpha-level and is only honored by servers that enable the TTLAfterFinished feature. - withTtlSecondsAfterFinished(ttlSecondsAfterFinished):: self + __specMixin({ ttlSecondsAfterFinished: ttlSecondsAfterFinished }), - }, - specType:: hidden.batch.v1.jobSpec, - }, - }, - }, - }, - certificates:: { - v1beta1:: { - local apiVersion = { apiVersion: 'certificates/v1beta1' }, - certificateSigningRequestCondition:: { - new():: {}, - // timestamp for the last update to this condition - withLastUpdateTime(lastUpdateTime):: self + { lastUpdateTime: lastUpdateTime }, - // human readable message with details about the request state - withMessage(message):: self + { message: message }, - // brief reason for the request state - withReason(reason):: self + { reason: reason }, - // request approval state, currently Approved or Denied. - withType(type):: self + { type: type }, - mixin:: {}, - }, - certificateSigningRequestList:: { - new(items=''):: self.withItems(items), - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.certificates.v1beta1.certificateSigningRequest, - mixin:: {}, - }, - // This information is immutable after the request is created. Only the Request and Usages fields can be set on creation, other fields are derived by Kubernetes and cannot be modified by users. - certificateSigningRequestSpec:: { - new():: {}, - // Extra information about the requesting user. See user.Info interface for details. - withExtra(extra):: self + { extra: extra }, - // Extra information about the requesting user. See user.Info interface for details. - withExtraMixin(extra):: self + { extra+: extra }, - // Group information about the requesting user. See user.Info interface for details. - withGroups(groups):: self + if std.type(groups) == 'array' then { groups: groups } else { groups: [groups] }, - // Group information about the requesting user. See user.Info interface for details. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then { groups+: groups } else { groups+: [groups] }, - // Base64-encoded PKCS#10 CSR data - withRequest(request):: self + { request: request }, - // UID information about the requesting user. See user.Info interface for details. - withUid(uid):: self + { uid: uid }, - // allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 - // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 - withUsages(usages):: self + if std.type(usages) == 'array' then { usages: usages } else { usages: [usages] }, - // allowedUsages specifies a set of usage contexts the key will be valid for. See: https://tools.ietf.org/html/rfc5280#section-4.2.1.3 - // https://tools.ietf.org/html/rfc5280#section-4.2.1.12 - withUsagesMixin(usages):: self + if std.type(usages) == 'array' then { usages+: usages } else { usages+: [usages] }, - // Information about the requesting user. See user.Info interface for details. - withUsername(username):: self + { username: username }, - mixin:: {}, - }, - certificateSigningRequestStatus:: { - new():: {}, - // If request was approved, the controller will place the issued certificate here. - withCertificate(certificate):: self + { certificate: certificate }, - // Conditions applied to the request, such as approval or denial. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Conditions applied to the request, such as approval or denial. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.certificates.v1beta1.certificateSigningRequestCondition, - mixin:: {}, - }, - }, - }, - coordination:: { - v1:: { - local apiVersion = { apiVersion: 'coordination/v1' }, - // LeaseList is a list of Lease objects. - leaseList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.coordination.v1.lease, - mixin:: {}, - }, - // LeaseSpec is a specification of a Lease. - leaseSpec:: { - new():: {}, - // acquireTime is a time when the current lease was acquired. - withAcquireTime(acquireTime):: self + { acquireTime: acquireTime }, - // holderIdentity contains the identity of the holder of a current lease. - withHolderIdentity(holderIdentity):: self + { holderIdentity: holderIdentity }, - // leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime. - withLeaseDurationSeconds(leaseDurationSeconds):: self + { leaseDurationSeconds: leaseDurationSeconds }, - // leaseTransitions is the number of transitions of a lease between holders. - withLeaseTransitions(leaseTransitions):: self + { leaseTransitions: leaseTransitions }, - // renewTime is a time when the current holder of a lease has last updated the lease. - withRenewTime(renewTime):: self + { renewTime: renewTime }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'coordination/v1beta1' }, - // LeaseList is a list of Lease objects. - leaseList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.coordination.v1beta1.lease, - mixin:: {}, - }, - // LeaseSpec is a specification of a Lease. - leaseSpec:: { - new():: {}, - // acquireTime is a time when the current lease was acquired. - withAcquireTime(acquireTime):: self + { acquireTime: acquireTime }, - // holderIdentity contains the identity of the holder of a current lease. - withHolderIdentity(holderIdentity):: self + { holderIdentity: holderIdentity }, - // leaseDurationSeconds is a duration that candidates for a lease need to wait to force acquire it. This is measure against time of last observed RenewTime. - withLeaseDurationSeconds(leaseDurationSeconds):: self + { leaseDurationSeconds: leaseDurationSeconds }, - // leaseTransitions is the number of transitions of a lease between holders. - withLeaseTransitions(leaseTransitions):: self + { leaseTransitions: leaseTransitions }, - // renewTime is a time when the current holder of a lease has last updated the lease. - withRenewTime(renewTime):: self + { renewTime: renewTime }, - mixin:: {}, - }, - }, - }, - core:: { - intstr:: { - local apiVersion = { apiVersion: 'intstr' }, - // IntOrString is a type that can hold an int32 or a string. When used in JSON or YAML marshalling and unmarshalling, it produces or consumes the inner type. This allows you to have, for example, a JSON field that can accept a name or number. - intOrString:: { - new():: {}, - mixin:: {}, - }, - }, - resource:: { - local apiVersion = { apiVersion: 'resource' }, - // Quantity is a fixed-point representation of a number. It provides convenient marshaling/unmarshaling in JSON and YAML, in addition to String() and Int64() accessors. - // - // The serialization format is: - // - // ::= - // (Note that may be empty, from the "" case in .) - // ::= 0 | 1 | ... | 9 ::= | ::= | . | . | . ::= "+" | "-" ::= | ::= | | ::= Ki | Mi | Gi | Ti | Pi | Ei - // (International System of units; See: http://physics.nist.gov/cuu/Units/binary.html) - // ::= m | "" | k | M | G | T | P | E - // (Note that 1024 = 1Ki but 1000 = 1k; I didn't choose the capitalization.) - // ::= "e" | "E" - // - // No matter which of the three exponent forms is used, no quantity may represent a number greater than 2^63-1 in magnitude, nor may it have more than 3 decimal places. Numbers larger or more precise will be capped or rounded up. (E.g.: 0.1m will rounded up to 1m.) This may be extended in the future if we require larger or smaller quantities. - // - // When a Quantity is parsed from a string, it will remember the type of suffix it had, and will use the same type again when it is serialized. - // - // Before serializing, Quantity will be put in "canonical form". This means that Exponent/suffix will be adjusted up or down (with a corresponding increase or decrease in Mantissa) such that: - // a. No precision is lost - // b. No fractional digits will be emitted - // c. The exponent (or suffix) is as large as possible. - // The sign will be omitted unless the number is negative. - // - // Examples: - // 1.5 will be serialized as "1500m" - // 1.5Gi will be serialized as "1536Mi" - // - // Note that the quantity will NEVER be internally represented by a floating point number. That is the whole point of this exercise. - // - // Non-canonical values will still parse as long as they are well formed, but will be re-emitted in their canonical form. (So always use canonical form, or don't diff.) - // - // This format is intended to make it difficult to use these numbers without writing some sort of special handling code in the hopes that that will cause implementors to also use a fixed point implementation. - quantity:: { - new():: {}, - mixin:: {}, - }, - }, - v1:: { - local apiVersion = { apiVersion: 'v1' }, - // Affinity is a group of affinity scheduling rules. - affinity:: { - new():: {}, - mixin:: { - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = { nodeAffinity+: nodeAffinity }, - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = { podAffinity+: podAffinity }, - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = { podAntiAffinity+: podAntiAffinity }, - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - }, - // AttachedVolume describes a volume attached to a node - attachedVolume:: { - new():: {}, - // DevicePath represents the device path where the volume should be available - withDevicePath(devicePath):: self + { devicePath: devicePath }, - // Name of the attached volume - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Represents a Persistent Disk resource in AWS. - // - // An AWS EBS disk must exist before mounting to a container. The disk must also be in the same AWS zone as the kubelet. An AWS EBS disk can only be mounted as read/write once. AWS EBS volumes support ownership management and SELinux relabeling. - awsElasticBlockStoreVolumeSource:: { - new():: {}, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withFsType(fsType):: self + { fsType: fsType }, - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). - withPartition(partition):: self + { partition: partition }, - // Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withVolumeId(volumeId):: self + { volumeID: volumeId }, - mixin:: {}, - }, - // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod. - azureDiskVolumeSource:: { - new():: {}, - // Host Caching mode: None, Read Only, Read Write. - withCachingMode(cachingMode):: self + { cachingMode: cachingMode }, - // The Name of the data disk in the blob storage - withDiskName(diskName):: self + { diskName: diskName }, - // The URI the data disk in the blob storage - withDiskUri(diskUri):: self + { diskURI: diskUri }, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // AzureFile represents an Azure File Service mount on the host and bind mount to the pod. - azureFilePersistentVolumeSource:: { - new():: {}, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // the name of secret that contains Azure Storage Account Name and Key - withSecretName(secretName):: self + { secretName: secretName }, - // the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod - withSecretNamespace(secretNamespace):: self + { secretNamespace: secretNamespace }, - // Share Name - withShareName(shareName):: self + { shareName: shareName }, - mixin:: {}, - }, - // AzureFile represents an Azure File Service mount on the host and bind mount to the pod. - azureFileVolumeSource:: { - new():: {}, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // the name of secret that contains Azure Storage Account Name and Key - withSecretName(secretName):: self + { secretName: secretName }, - // Share Name - withShareName(shareName):: self + { shareName: shareName }, - mixin:: {}, - }, - // Adds and removes POSIX capabilities from running containers. - capabilities:: { - new():: {}, - // Added capabilities - withAdd(add):: self + if std.type(add) == 'array' then { add: add } else { add: [add] }, - // Added capabilities - withAddMixin(add):: self + if std.type(add) == 'array' then { add+: add } else { add+: [add] }, - // Removed capabilities - withDrop(drop):: self + if std.type(drop) == 'array' then { drop: drop } else { drop: [drop] }, - // Removed capabilities - withDropMixin(drop):: self + if std.type(drop) == 'array' then { drop+: drop } else { drop+: [drop] }, - mixin:: {}, - }, - // Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling. - cephFsPersistentVolumeSource:: { - new():: {}, - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then { monitors: monitors } else { monitors: [monitors] }, - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then { monitors+: monitors } else { monitors+: [monitors] }, - // Optional: Used as the mounted root, rather than the full Ceph tree, default is / - withPath(path):: self + { path: path }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withSecretFile(secretFile):: self + { secretFile: secretFile }, - // Optional: User is the rados user name, default is admin More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withUser(user):: self + { user: user }, - mixin:: { - // Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // Represents a Ceph Filesystem mount that lasts the lifetime of a pod Cephfs volumes do not support ownership management or SELinux relabeling. - cephFsVolumeSource:: { - new():: {}, - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then { monitors: monitors } else { monitors: [monitors] }, - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then { monitors+: monitors } else { monitors+: [monitors] }, - // Optional: Used as the mounted root, rather than the full Ceph tree, default is / - withPath(path):: self + { path: path }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withSecretFile(secretFile):: self + { secretFile: secretFile }, - // Optional: User is the rados user name, default is admin More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withUser(user):: self + { user: user }, - mixin:: { - // Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling. - cinderPersistentVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withFsType(fsType):: self + { fsType: fsType }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // volume id used to identify the volume in cinder More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withVolumeId(volumeId):: self + { volumeID: volumeId }, - mixin:: { - // Optional: points to a secret object containing parameters used to connect to OpenStack. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // Represents a cinder volume resource in Openstack. A Cinder volume must exist before mounting to a container. The volume must also be in the same region as the kubelet. Cinder volumes support ownership management and SELinux relabeling. - cinderVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withFsType(fsType):: self + { fsType: fsType }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // volume id used to identify the volume in cinder More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withVolumeId(volumeId):: self + { volumeID: volumeId }, - mixin:: { - // Optional: points to a secret object containing parameters used to connect to OpenStack. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // ClientIPConfig represents the configurations of Client IP based session affinity. - clientIpConfig:: { - new():: {}, - // timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours). - withTimeoutSeconds(timeoutSeconds):: self + { timeoutSeconds: timeoutSeconds }, - mixin:: {}, - }, - // Information about the condition of a component. - componentCondition:: { - new():: {}, - // Condition error code for a component. For example, a health check error code. - withErrorParam(errorParam):: self + { "error": errorParam }, - // Message about the condition for a component. For example, information about a health check. - withMessage(message):: self + { message: message }, - // Status of the condition for a component. Valid values for "Healthy": "True", "False", or "Unknown". - withStatus(status):: self + { status: status }, - // Type of condition for a component. Valid value: "Healthy" - withType(type):: self + { type: type }, - mixin:: {}, - }, - // ComponentStatus (and ComponentStatusList) holds the cluster validation info. - componentStatus:: { - new():: {}, - // List of component conditions observed - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // List of component conditions observed - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.core.v1.componentCondition, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - }, - }, - // Status of all the conditions for the component as a list of ComponentStatus objects. - componentStatusList:: { - new():: {}, - // List of ComponentStatus objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ComponentStatus objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.componentStatus, - mixin:: {}, - }, - // ConfigMapEnvSource selects a ConfigMap to populate the environment variables with. - // - // The contents of the target ConfigMap's Data field will represent the key-value pairs as environment variables. - configMapEnvSource:: { - new():: {}, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the ConfigMap must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // Selects a key from a ConfigMap. - configMapKeySelector:: { - new():: {}, - // The key to select. - withKey(key):: self + { key: key }, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the ConfigMap or it's key must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // ConfigMapList is a resource containing a list of ConfigMap objects. - configMapList:: { - new(items=''):: self.withItems(items), - // Items is the list of ConfigMaps. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of ConfigMaps. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.configMap, - mixin:: {}, - }, - // ConfigMapNodeConfigSource contains the information to reference a ConfigMap as a config source for the Node. - configMapNodeConfigSource:: { - new():: {}, - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + { kubeletConfigKey: kubeletConfigKey }, - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + { name: name }, - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + { namespace: namespace }, - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + { resourceVersion: resourceVersion }, - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + { uid: uid }, - mixin:: {}, - }, - // Adapts a ConfigMap into a projected volume. - // - // The contents of the target ConfigMap's Data field will be presented in a projected volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. Note that this is identical to a configmap volume source without the default mode. - configMapProjection:: { - new():: {}, - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the ConfigMap or it's keys must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // Adapts a ConfigMap into a volume. - // - // The contents of the target ConfigMap's Data field will be presented in a volume as files using the keys in the Data field as the file names, unless the items element is populated with specific mappings of keys to paths. ConfigMap volumes support ownership management and SELinux relabeling. - configMapVolumeSource:: { - new():: {}, - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + { defaultMode: defaultMode }, - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the ConfigMap or it's keys must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // A single application container that you want to run within a pod. - container:: { - new(name='', image=''):: self.withImage(image).withName(name), - // Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell - withArgs(args):: self + if std.type(args) == 'array' then { args: args } else { args: [args] }, - // Arguments to the entrypoint. The docker image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell - withArgsMixin(args):: self + if std.type(args) == 'array' then { args+: args } else { args+: [args] }, - // Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell - withCommand(command):: self + if std.type(command) == 'array' then { command: command } else { command: [command] }, - // Entrypoint array. Not executed within a shell. The docker image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell - withCommandMixin(command):: self + if std.type(command) == 'array' then { command+: command } else { command+: [command] }, - // List of environment variables to set in the container. Cannot be updated. - withEnv(env):: self + if std.type(env) == 'array' then { env: env } else { env: [env] }, - // List of environment variables to set in the container. Cannot be updated. - withEnvMixin(env):: self + if std.type(env) == 'array' then { env+: env } else { env+: [env] }, - envType:: hidden.core.v1.envVar, - // List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated. - withEnvFrom(envFrom):: self + if std.type(envFrom) == 'array' then { envFrom: envFrom } else { envFrom: [envFrom] }, - // List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated. - withEnvFromMixin(envFrom):: self + if std.type(envFrom) == 'array' then { envFrom+: envFrom } else { envFrom+: [envFrom] }, - envFromType:: hidden.core.v1.envFromSource, - // Docker image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets. - withImage(image):: self + { image: image }, - // Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images - withImagePullPolicy(imagePullPolicy):: self + { imagePullPolicy: imagePullPolicy }, - // Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated. - withName(name):: self + { name: name }, - // List of ports to expose from the container. Exposing a port here gives the system additional information about the network connections a container uses, but is primarily informational. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default "0.0.0.0" address inside a container will be accessible from the network. Cannot be updated. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // List of ports to expose from the container. Exposing a port here gives the system additional information about the network connections a container uses, but is primarily informational. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default "0.0.0.0" address inside a container will be accessible from the network. Cannot be updated. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.core.v1.containerPort, - // Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false. - withStdin(stdin):: self + { stdin: stdin }, - // Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false - withStdinOnce(stdinOnce):: self + { stdinOnce: stdinOnce }, - // Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated. - withTerminationMessagePath(terminationMessagePath):: self + { terminationMessagePath: terminationMessagePath }, - // Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated. - withTerminationMessagePolicy(terminationMessagePolicy):: self + { terminationMessagePolicy: terminationMessagePolicy }, - // Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false. - withTty(tty):: self + { tty: tty }, - // volumeDevices is the list of block devices to be used by the container. This is a beta feature. - withVolumeDevices(volumeDevices):: self + if std.type(volumeDevices) == 'array' then { volumeDevices: volumeDevices } else { volumeDevices: [volumeDevices] }, - // volumeDevices is the list of block devices to be used by the container. This is a beta feature. - withVolumeDevicesMixin(volumeDevices):: self + if std.type(volumeDevices) == 'array' then { volumeDevices+: volumeDevices } else { volumeDevices+: [volumeDevices] }, - volumeDevicesType:: hidden.core.v1.volumeDevice, - // Pod volumes to mount into the container's filesystem. Cannot be updated. - withVolumeMounts(volumeMounts):: self + if std.type(volumeMounts) == 'array' then { volumeMounts: volumeMounts } else { volumeMounts: [volumeMounts] }, - // Pod volumes to mount into the container's filesystem. Cannot be updated. - withVolumeMountsMixin(volumeMounts):: self + if std.type(volumeMounts) == 'array' then { volumeMounts+: volumeMounts } else { volumeMounts+: [volumeMounts] }, - volumeMountsType:: hidden.core.v1.volumeMount, - // Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated. - withWorkingDir(workingDir):: self + { workingDir: workingDir }, - mixin:: { - // Actions that the management system should take in response to container lifecycle events. Cannot be updated. - lifecycle:: { - local __lifecycleMixin(lifecycle) = { lifecycle+: lifecycle }, - mixinInstance(lifecycle):: __lifecycleMixin(lifecycle), - // PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks - postStart:: { - local __postStartMixin(postStart) = __lifecycleMixin({ postStart+: postStart }), - mixinInstance(postStart):: __postStartMixin(postStart), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __postStartMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __postStartMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __postStartMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - postStartType:: hidden.core.v1.handler, - // PreStop is called immediately before a container is terminated due to an API request or management event such as liveness probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The reason for termination is passed to the handler. The Pod's termination grace period countdown begins before the PreStop hooked is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period. Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks - preStop:: { - local __preStopMixin(preStop) = __lifecycleMixin({ preStop+: preStop }), - mixinInstance(preStop):: __preStopMixin(preStop), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __preStopMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __preStopMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __preStopMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - preStopType:: hidden.core.v1.handler, - }, - lifecycleType:: hidden.core.v1.lifecycle, - // Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - livenessProbe:: { - local __livenessProbeMixin(livenessProbe) = { livenessProbe+: livenessProbe }, - mixinInstance(livenessProbe):: __livenessProbeMixin(livenessProbe), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __livenessProbeMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1. - withFailureThreshold(failureThreshold):: self + __livenessProbeMixin({ failureThreshold: failureThreshold }), - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __livenessProbeMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withInitialDelaySeconds(initialDelaySeconds):: self + __livenessProbeMixin({ initialDelaySeconds: initialDelaySeconds }), - // How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1. - withPeriodSeconds(periodSeconds):: self + __livenessProbeMixin({ periodSeconds: periodSeconds }), - // Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness. Minimum value is 1. - withSuccessThreshold(successThreshold):: self + __livenessProbeMixin({ successThreshold: successThreshold }), - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __livenessProbeMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - // Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withTimeoutSeconds(timeoutSeconds):: self + __livenessProbeMixin({ timeoutSeconds: timeoutSeconds }), - }, - livenessProbeType:: hidden.core.v1.probe, - // Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - readinessProbe:: { - local __readinessProbeMixin(readinessProbe) = { readinessProbe+: readinessProbe }, - mixinInstance(readinessProbe):: __readinessProbeMixin(readinessProbe), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __readinessProbeMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1. - withFailureThreshold(failureThreshold):: self + __readinessProbeMixin({ failureThreshold: failureThreshold }), - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __readinessProbeMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withInitialDelaySeconds(initialDelaySeconds):: self + __readinessProbeMixin({ initialDelaySeconds: initialDelaySeconds }), - // How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1. - withPeriodSeconds(periodSeconds):: self + __readinessProbeMixin({ periodSeconds: periodSeconds }), - // Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness. Minimum value is 1. - withSuccessThreshold(successThreshold):: self + __readinessProbeMixin({ successThreshold: successThreshold }), - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __readinessProbeMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - // Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withTimeoutSeconds(timeoutSeconds):: self + __readinessProbeMixin({ timeoutSeconds: timeoutSeconds }), - }, - readinessProbeType:: hidden.core.v1.probe, - // Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - resources:: { - local __resourcesMixin(resources) = { resources+: resources }, - mixinInstance(resources):: __resourcesMixin(resources), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimits(limits):: self + __resourcesMixin({ limits: limits }), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimitsMixin(limits):: self + __resourcesMixin({ limits+: limits }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequests(requests):: self + __resourcesMixin({ requests: requests }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequestsMixin(requests):: self + __resourcesMixin({ requests+: requests }), - }, - resourcesType:: hidden.core.v1.resourceRequirements, - // Security options the pod should run with. More info: https://kubernetes.io/docs/concepts/policy/security-context/ More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - securityContext:: { - local __securityContextMixin(securityContext) = { securityContext+: securityContext }, - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + __securityContextMixin({ allowPrivilegeEscalation: allowPrivilegeEscalation }), - // The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. - capabilities:: { - local __capabilitiesMixin(capabilities) = __securityContextMixin({ capabilities+: capabilities }), - mixinInstance(capabilities):: __capabilitiesMixin(capabilities), - // Added capabilities - withAdd(add):: self + if std.type(add) == 'array' then __capabilitiesMixin({ add: add }) else __capabilitiesMixin({ add: [add] }), - // Added capabilities - withAddMixin(add):: self + if std.type(add) == 'array' then __capabilitiesMixin({ add+: add }) else __capabilitiesMixin({ add+: [add] }), - // Removed capabilities - withDrop(drop):: self + if std.type(drop) == 'array' then __capabilitiesMixin({ drop: drop }) else __capabilitiesMixin({ drop: [drop] }), - // Removed capabilities - withDropMixin(drop):: self + if std.type(drop) == 'array' then __capabilitiesMixin({ drop+: drop }) else __capabilitiesMixin({ drop+: [drop] }), - }, - capabilitiesType:: hidden.core.v1.capabilities, - // Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. - withPrivileged(privileged):: self + __securityContextMixin({ privileged: privileged }), - // procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. - withProcMount(procMount):: self + __securityContextMixin({ procMount: procMount }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - securityContextType:: hidden.core.v1.securityContext, - }, - }, - // Describe a container image - containerImage:: { - new():: {}, - // Names by which this image is known. e.g. ["k8s.gcr.io/hyperkube:v1.0.7", "dockerhub.io/google_containers/hyperkube:v1.0.7"] - withNames(names):: self + if std.type(names) == 'array' then { names: names } else { names: [names] }, - // Names by which this image is known. e.g. ["k8s.gcr.io/hyperkube:v1.0.7", "dockerhub.io/google_containers/hyperkube:v1.0.7"] - withNamesMixin(names):: self + if std.type(names) == 'array' then { names+: names } else { names+: [names] }, - // The size of the image in bytes. - withSizeBytes(sizeBytes):: self + { sizeBytes: sizeBytes }, - mixin:: {}, - }, - // ContainerPort represents a network port in a single container. - containerPort:: { - new(containerPort=''):: self.withContainerPort(containerPort), - newNamed(containerPort='', name=''):: self.withContainerPort(containerPort).withName(name), - // Number of port to expose on the pod's IP address. This must be a valid port number, 0 < x < 65536. - withContainerPort(containerPort):: self + { containerPort: containerPort }, - // What host IP to bind the external port to. - withHostIp(hostIp):: self + { hostIP: hostIp }, - // Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this. - withHostPort(hostPort):: self + { hostPort: hostPort }, - // If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services. - withName(name):: self + { name: name }, - // Protocol for port. Must be UDP, TCP, or SCTP. Defaults to "TCP". - withProtocol(protocol):: self + { protocol: protocol }, - mixin:: {}, - }, - // ContainerState holds a possible state of container. Only one of its members may be specified. If none of them is specified, the default one is ContainerStateWaiting. - containerState:: { - new():: {}, - mixin:: { - // Details about a running container - running:: { - local __runningMixin(running) = { running+: running }, - mixinInstance(running):: __runningMixin(running), - // Time at which the container was last (re-)started - withStartedAt(startedAt):: self + __runningMixin({ startedAt: startedAt }), - }, - runningType:: hidden.core.v1.containerStateRunning, - // Details about a terminated container - terminated:: { - local __terminatedMixin(terminated) = { terminated+: terminated }, - mixinInstance(terminated):: __terminatedMixin(terminated), - // Container's ID in the format 'docker://' - withContainerId(containerId):: self + __terminatedMixin({ containerID: containerId }), - // Exit status from the last termination of the container - withExitCode(exitCode):: self + __terminatedMixin({ exitCode: exitCode }), - // Time at which the container last terminated - withFinishedAt(finishedAt):: self + __terminatedMixin({ finishedAt: finishedAt }), - // Message regarding the last termination of the container - withMessage(message):: self + __terminatedMixin({ message: message }), - // (brief) reason from the last termination of the container - withReason(reason):: self + __terminatedMixin({ reason: reason }), - // Signal from the last termination of the container - withSignal(signal):: self + __terminatedMixin({ signal: signal }), - // Time at which previous execution of the container started - withStartedAt(startedAt):: self + __terminatedMixin({ startedAt: startedAt }), - }, - terminatedType:: hidden.core.v1.containerStateTerminated, - // Details about a waiting container - waiting:: { - local __waitingMixin(waiting) = { waiting+: waiting }, - mixinInstance(waiting):: __waitingMixin(waiting), - // Message regarding why the container is not yet running. - withMessage(message):: self + __waitingMixin({ message: message }), - // (brief) reason the container is not yet running. - withReason(reason):: self + __waitingMixin({ reason: reason }), - }, - waitingType:: hidden.core.v1.containerStateWaiting, - }, - }, - // ContainerStateRunning is a running state of a container. - containerStateRunning:: { - new():: {}, - // Time at which the container was last (re-)started - withStartedAt(startedAt):: self + { startedAt: startedAt }, - mixin:: {}, - }, - // ContainerStateTerminated is a terminated state of a container. - containerStateTerminated:: { - new():: {}, - // Container's ID in the format 'docker://' - withContainerId(containerId):: self + { containerID: containerId }, - // Exit status from the last termination of the container - withExitCode(exitCode):: self + { exitCode: exitCode }, - // Time at which the container last terminated - withFinishedAt(finishedAt):: self + { finishedAt: finishedAt }, - // Message regarding the last termination of the container - withMessage(message):: self + { message: message }, - // (brief) reason from the last termination of the container - withReason(reason):: self + { reason: reason }, - // Signal from the last termination of the container - withSignal(signal):: self + { signal: signal }, - // Time at which previous execution of the container started - withStartedAt(startedAt):: self + { startedAt: startedAt }, - mixin:: {}, - }, - // ContainerStateWaiting is a waiting state of a container. - containerStateWaiting:: { - new():: {}, - // Message regarding why the container is not yet running. - withMessage(message):: self + { message: message }, - // (brief) reason the container is not yet running. - withReason(reason):: self + { reason: reason }, - mixin:: {}, - }, - // ContainerStatus contains details for the current status of this container. - containerStatus:: { - new():: {}, - // Container's ID in the format 'docker://'. - withContainerId(containerId):: self + { containerID: containerId }, - // The image the container is running. More info: https://kubernetes.io/docs/concepts/containers/images - withImage(image):: self + { image: image }, - // ImageID of the container's image. - withImageId(imageId):: self + { imageID: imageId }, - // This must be a DNS_LABEL. Each container in a pod must have a unique name. Cannot be updated. - withName(name):: self + { name: name }, - // Specifies whether the container has passed its readiness probe. - withReady(ready):: self + { ready: ready }, - // The number of times the container has been restarted, currently based on the number of dead containers that have not yet been removed. Note that this is calculated from dead containers. But those containers are subject to garbage collection. This value will get capped at 5 by GC. - withRestartCount(restartCount):: self + { restartCount: restartCount }, - mixin:: { - // Details about the container's last termination condition. - lastState:: { - local __lastStateMixin(lastState) = { lastState+: lastState }, - mixinInstance(lastState):: __lastStateMixin(lastState), - // Details about a running container - running:: { - local __runningMixin(running) = __lastStateMixin({ running+: running }), - mixinInstance(running):: __runningMixin(running), - // Time at which the container was last (re-)started - withStartedAt(startedAt):: self + __runningMixin({ startedAt: startedAt }), - }, - runningType:: hidden.core.v1.containerStateRunning, - // Details about a terminated container - terminated:: { - local __terminatedMixin(terminated) = __lastStateMixin({ terminated+: terminated }), - mixinInstance(terminated):: __terminatedMixin(terminated), - // Container's ID in the format 'docker://' - withContainerId(containerId):: self + __terminatedMixin({ containerID: containerId }), - // Exit status from the last termination of the container - withExitCode(exitCode):: self + __terminatedMixin({ exitCode: exitCode }), - // Time at which the container last terminated - withFinishedAt(finishedAt):: self + __terminatedMixin({ finishedAt: finishedAt }), - // Message regarding the last termination of the container - withMessage(message):: self + __terminatedMixin({ message: message }), - // (brief) reason from the last termination of the container - withReason(reason):: self + __terminatedMixin({ reason: reason }), - // Signal from the last termination of the container - withSignal(signal):: self + __terminatedMixin({ signal: signal }), - // Time at which previous execution of the container started - withStartedAt(startedAt):: self + __terminatedMixin({ startedAt: startedAt }), - }, - terminatedType:: hidden.core.v1.containerStateTerminated, - // Details about a waiting container - waiting:: { - local __waitingMixin(waiting) = __lastStateMixin({ waiting+: waiting }), - mixinInstance(waiting):: __waitingMixin(waiting), - // Message regarding why the container is not yet running. - withMessage(message):: self + __waitingMixin({ message: message }), - // (brief) reason the container is not yet running. - withReason(reason):: self + __waitingMixin({ reason: reason }), - }, - waitingType:: hidden.core.v1.containerStateWaiting, - }, - lastStateType:: hidden.core.v1.containerState, - // Details about the container's current condition. - state:: { - local __stateMixin(state) = { state+: state }, - mixinInstance(state):: __stateMixin(state), - // Details about a running container - running:: { - local __runningMixin(running) = __stateMixin({ running+: running }), - mixinInstance(running):: __runningMixin(running), - // Time at which the container was last (re-)started - withStartedAt(startedAt):: self + __runningMixin({ startedAt: startedAt }), - }, - runningType:: hidden.core.v1.containerStateRunning, - // Details about a terminated container - terminated:: { - local __terminatedMixin(terminated) = __stateMixin({ terminated+: terminated }), - mixinInstance(terminated):: __terminatedMixin(terminated), - // Container's ID in the format 'docker://' - withContainerId(containerId):: self + __terminatedMixin({ containerID: containerId }), - // Exit status from the last termination of the container - withExitCode(exitCode):: self + __terminatedMixin({ exitCode: exitCode }), - // Time at which the container last terminated - withFinishedAt(finishedAt):: self + __terminatedMixin({ finishedAt: finishedAt }), - // Message regarding the last termination of the container - withMessage(message):: self + __terminatedMixin({ message: message }), - // (brief) reason from the last termination of the container - withReason(reason):: self + __terminatedMixin({ reason: reason }), - // Signal from the last termination of the container - withSignal(signal):: self + __terminatedMixin({ signal: signal }), - // Time at which previous execution of the container started - withStartedAt(startedAt):: self + __terminatedMixin({ startedAt: startedAt }), - }, - terminatedType:: hidden.core.v1.containerStateTerminated, - // Details about a waiting container - waiting:: { - local __waitingMixin(waiting) = __stateMixin({ waiting+: waiting }), - mixinInstance(waiting):: __waitingMixin(waiting), - // Message regarding why the container is not yet running. - withMessage(message):: self + __waitingMixin({ message: message }), - // (brief) reason the container is not yet running. - withReason(reason):: self + __waitingMixin({ reason: reason }), - }, - waitingType:: hidden.core.v1.containerStateWaiting, - }, - stateType:: hidden.core.v1.containerState, - }, - }, - // Represents storage that is managed by an external CSI volume driver (Beta feature) - csiPersistentVolumeSource:: { - new():: {}, - // Driver is the name of the driver to use for this volume. Required. - withDriver(driver):: self + { driver: driver }, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". - withFsType(fsType):: self + { fsType: fsType }, - // Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write). - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Attributes of the volume to publish. - withVolumeAttributes(volumeAttributes):: self + { volumeAttributes: volumeAttributes }, - // Attributes of the volume to publish. - withVolumeAttributesMixin(volumeAttributes):: self + { volumeAttributes+: volumeAttributes }, - // VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required. - withVolumeHandle(volumeHandle):: self + { volumeHandle: volumeHandle }, - mixin:: { - // ControllerPublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI ControllerPublishVolume and ControllerUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - controllerPublishSecretRef:: { - local __controllerPublishSecretRefMixin(controllerPublishSecretRef) = { controllerPublishSecretRef+: controllerPublishSecretRef }, - mixinInstance(controllerPublishSecretRef):: __controllerPublishSecretRefMixin(controllerPublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __controllerPublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __controllerPublishSecretRefMixin({ namespace: namespace }), - }, - controllerPublishSecretRefType:: hidden.core.v1.secretReference, - // NodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodePublishSecretRef:: { - local __nodePublishSecretRefMixin(nodePublishSecretRef) = { nodePublishSecretRef+: nodePublishSecretRef }, - mixinInstance(nodePublishSecretRef):: __nodePublishSecretRefMixin(nodePublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodePublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodePublishSecretRefMixin({ namespace: namespace }), - }, - nodePublishSecretRefType:: hidden.core.v1.secretReference, - // NodeStageSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodeStageVolume and NodeStageVolume and NodeUnstageVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodeStageSecretRef:: { - local __nodeStageSecretRefMixin(nodeStageSecretRef) = { nodeStageSecretRef+: nodeStageSecretRef }, - mixinInstance(nodeStageSecretRef):: __nodeStageSecretRefMixin(nodeStageSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodeStageSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodeStageSecretRefMixin({ namespace: namespace }), - }, - nodeStageSecretRefType:: hidden.core.v1.secretReference, - }, - }, - // Represents a source location of a volume to mount, managed by an external CSI driver - csiVolumeSource:: { - new():: {}, - // Driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster. - withDriver(driver):: self + { driver: driver }, - // Filesystem type to mount. Ex. "ext4", "xfs", "ntfs". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply. - withFsType(fsType):: self + { fsType: fsType }, - // Specifies a read-only configuration for the volume. Defaults to false (read/write). - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values. - withVolumeAttributes(volumeAttributes):: self + { volumeAttributes: volumeAttributes }, - // VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values. - withVolumeAttributesMixin(volumeAttributes):: self + { volumeAttributes+: volumeAttributes }, - mixin:: { - // NodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed. - nodePublishSecretRef:: { - local __nodePublishSecretRefMixin(nodePublishSecretRef) = { nodePublishSecretRef+: nodePublishSecretRef }, - mixinInstance(nodePublishSecretRef):: __nodePublishSecretRefMixin(nodePublishSecretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __nodePublishSecretRefMixin({ name: name }), - }, - nodePublishSecretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // DaemonEndpoint contains information about a single Daemon endpoint. - daemonEndpoint:: { - new():: {}, - // Port number of the given endpoint. - withPort(port):: self + { Port: port }, - mixin:: {}, - }, - // Represents downward API info for projecting into a projected volume. Note that this is identical to a downwardAPI volume source without the default mode. - downwardApiProjection:: { - new():: {}, - // Items is a list of DownwardAPIVolume file - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of DownwardAPIVolume file - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.downwardApiVolumeFile, - mixin:: {}, - }, - // DownwardAPIVolumeFile represents information to create the file containing the pod field - downwardApiVolumeFile:: { - new():: {}, - // Optional: mode bits to use on this file, must be a value between 0 and 0777. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withMode(mode):: self + { mode: mode }, - // Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..' - withPath(path):: self + { path: path }, - mixin:: { - // Required: Selects a field of the pod: only annotations, labels, name and namespace are supported. - fieldRef:: { - local __fieldRefMixin(fieldRef) = { fieldRef+: fieldRef }, - mixinInstance(fieldRef):: __fieldRefMixin(fieldRef), - // Path of the field to select in the specified API version. - withFieldPath(fieldPath):: self + __fieldRefMixin({ fieldPath: fieldPath }), - }, - fieldRefType:: hidden.core.v1.objectFieldSelector, - // Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported. - resourceFieldRef:: { - local __resourceFieldRefMixin(resourceFieldRef) = { resourceFieldRef+: resourceFieldRef }, - mixinInstance(resourceFieldRef):: __resourceFieldRefMixin(resourceFieldRef), - // Container name: required for volumes, optional for env vars - withContainerName(containerName):: self + __resourceFieldRefMixin({ containerName: containerName }), - // Specifies the output format of the exposed resources, defaults to "1" - divisor:: { - local __divisorMixin(divisor) = __resourceFieldRefMixin({ divisor+: divisor }), - mixinInstance(divisor):: __divisorMixin(divisor), - }, - divisorType:: hidden.core.resource.quantity, - // Required: resource to select - withResource(resource):: self + __resourceFieldRefMixin({ resource: resource }), - }, - resourceFieldRefType:: hidden.core.v1.resourceFieldSelector, - }, - }, - // DownwardAPIVolumeSource represents a volume containing downward API info. Downward API volumes support ownership management and SELinux relabeling. - downwardApiVolumeSource:: { - new():: {}, - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + { defaultMode: defaultMode }, - // Items is a list of downward API volume file - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of downward API volume file - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.downwardApiVolumeFile, - mixin:: {}, - }, - // Represents an empty directory for a pod. Empty directory volumes support ownership management and SELinux relabeling. - emptyDirVolumeSource:: { - new():: {}, - // What type of storage medium should back this directory. The default is "" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir - withMedium(medium):: self + { medium: medium }, - mixin:: { - // Total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: http://kubernetes.io/docs/user-guide/volumes#emptydir - sizeLimit:: { - local __sizeLimitMixin(sizeLimit) = { sizeLimit+: sizeLimit }, - mixinInstance(sizeLimit):: __sizeLimitMixin(sizeLimit), - }, - sizeLimitType:: hidden.core.resource.quantity, - }, - }, - // EndpointAddress is a tuple that describes single IP address. - endpointAddress:: { - new():: {}, - // The Hostname of this endpoint - withHostname(hostname):: self + { hostname: hostname }, - // The IP of this endpoint. May not be loopback (127.0.0.0/8), link-local (169.254.0.0/16), or link-local multicast ((224.0.0.0/24). IPv6 is also accepted but not fully supported on all platforms. Also, certain kubernetes components, like kube-proxy, are not IPv6 ready. - withIp(ip):: self + { ip: ip }, - // Optional: Node hosting this endpoint. This can be used to determine endpoints local to a node. - withNodeName(nodeName):: self + { nodeName: nodeName }, - mixin:: { - // Reference to object providing the endpoint. - targetRef:: { - local __targetRefMixin(targetRef) = { targetRef+: targetRef }, - mixinInstance(targetRef):: __targetRefMixin(targetRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __targetRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __targetRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __targetRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __targetRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __targetRefMixin({ uid: uid }), - }, - targetRefType:: hidden.core.v1.objectReference, - }, - }, - // EndpointPort is a tuple that describes a single port. - endpointPort:: { - new():: {}, - // The name of this port (corresponds to ServicePort.Name). Must be a DNS_LABEL. Optional only if one port is defined. - withName(name):: self + { name: name }, - // The port number of the endpoint. - withPort(port):: self + { port: port }, - // The IP protocol for this port. Must be UDP, TCP, or SCTP. Default is TCP. - withProtocol(protocol):: self + { protocol: protocol }, - mixin:: {}, - }, - // EndpointSubset is a group of addresses with a common set of ports. The expanded set of endpoints is the Cartesian product of Addresses x Ports. For example, given: - // { - // Addresses: [{"ip": "10.10.1.1"}, {"ip": "10.10.2.2"}], - // Ports: [{"name": "a", "port": 8675}, {"name": "b", "port": 309}] - // } - // The resulting set of endpoints can be viewed as: - // a: [ 10.10.1.1:8675, 10.10.2.2:8675 ], - // b: [ 10.10.1.1:309, 10.10.2.2:309 ] - endpointSubset:: { - new():: {}, - // IP addresses which offer the related ports that are marked as ready. These endpoints should be considered safe for load balancers and clients to utilize. - withAddresses(addresses):: self + if std.type(addresses) == 'array' then { addresses: addresses } else { addresses: [addresses] }, - // IP addresses which offer the related ports that are marked as ready. These endpoints should be considered safe for load balancers and clients to utilize. - withAddressesMixin(addresses):: self + if std.type(addresses) == 'array' then { addresses+: addresses } else { addresses+: [addresses] }, - addressesType:: hidden.core.v1.endpointAddress, - // IP addresses which offer the related ports but are not currently marked as ready because they have not yet finished starting, have recently failed a readiness check, or have recently failed a liveness check. - withNotReadyAddresses(notReadyAddresses):: self + if std.type(notReadyAddresses) == 'array' then { notReadyAddresses: notReadyAddresses } else { notReadyAddresses: [notReadyAddresses] }, - // IP addresses which offer the related ports but are not currently marked as ready because they have not yet finished starting, have recently failed a readiness check, or have recently failed a liveness check. - withNotReadyAddressesMixin(notReadyAddresses):: self + if std.type(notReadyAddresses) == 'array' then { notReadyAddresses+: notReadyAddresses } else { notReadyAddresses+: [notReadyAddresses] }, - notReadyAddressesType:: hidden.core.v1.endpointAddress, - // Port numbers available on the related IP addresses. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // Port numbers available on the related IP addresses. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.core.v1.endpointPort, - mixin:: {}, - }, - // EndpointsList is a list of endpoints. - endpointsList:: { - new(items=''):: self.withItems(items), - // List of endpoints. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of endpoints. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.endpoints, - mixin:: {}, - }, - // EnvFromSource represents the source of a set of ConfigMaps - envFromSource:: { - new():: {}, - // An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER. - withPrefix(prefix):: self + { prefix: prefix }, - mixin:: { - // The ConfigMap to select from - configMapRef:: { - local __configMapRefMixin(configMapRef) = { configMapRef+: configMapRef }, - mixinInstance(configMapRef):: __configMapRefMixin(configMapRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __configMapRefMixin({ name: name }), - // Specify whether the ConfigMap must be defined - withOptional(optional):: self + __configMapRefMixin({ optional: optional }), - }, - configMapRefType:: hidden.core.v1.configMapEnvSource, - // The Secret to select from - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - // Specify whether the Secret must be defined - withOptional(optional):: self + __secretRefMixin({ optional: optional }), - }, - secretRefType:: hidden.core.v1.secretEnvSource, - }, - }, - // EnvVar represents an environment variable present in a Container. - envVar:: { - new(name='', value=''):: self.withName(name).withValue(value), - fromSecretRef(name='', secretRefName='', secretRefKey=''):: self.withName(name) + self.mixin.valueFrom.secretKeyRef.withKey(secretRefKey).withName(secretRefName), - fromFieldPath(name='', fieldPath=''):: self.withName(name) + self.mixin.valueFrom.fieldRef.withFieldPath(fieldPath), - // Name of the environment variable. Must be a C_IDENTIFIER. - withName(name):: self + { name: name }, - // Variable references $(VAR_NAME) are expanded using the previous defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to "". - withValue(value):: self + { value: value }, - mixin:: { - // Source for the environment variable's value. Cannot be used if value is not empty. - valueFrom:: { - local __valueFromMixin(valueFrom) = { valueFrom+: valueFrom }, - mixinInstance(valueFrom):: __valueFromMixin(valueFrom), - // Selects a key of a ConfigMap. - configMapKeyRef:: { - local __configMapKeyRefMixin(configMapKeyRef) = __valueFromMixin({ configMapKeyRef+: configMapKeyRef }), - mixinInstance(configMapKeyRef):: __configMapKeyRefMixin(configMapKeyRef), - // The key to select. - withKey(key):: self + __configMapKeyRefMixin({ key: key }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __configMapKeyRefMixin({ name: name }), - // Specify whether the ConfigMap or it's key must be defined - withOptional(optional):: self + __configMapKeyRefMixin({ optional: optional }), - }, - configMapKeyRefType:: hidden.core.v1.configMapKeySelector, - // Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP. - fieldRef:: { - local __fieldRefMixin(fieldRef) = __valueFromMixin({ fieldRef+: fieldRef }), - mixinInstance(fieldRef):: __fieldRefMixin(fieldRef), - // Path of the field to select in the specified API version. - withFieldPath(fieldPath):: self + __fieldRefMixin({ fieldPath: fieldPath }), - }, - fieldRefType:: hidden.core.v1.objectFieldSelector, - // Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported. - resourceFieldRef:: { - local __resourceFieldRefMixin(resourceFieldRef) = __valueFromMixin({ resourceFieldRef+: resourceFieldRef }), - mixinInstance(resourceFieldRef):: __resourceFieldRefMixin(resourceFieldRef), - // Container name: required for volumes, optional for env vars - withContainerName(containerName):: self + __resourceFieldRefMixin({ containerName: containerName }), - // Specifies the output format of the exposed resources, defaults to "1" - divisor:: { - local __divisorMixin(divisor) = __resourceFieldRefMixin({ divisor+: divisor }), - mixinInstance(divisor):: __divisorMixin(divisor), - }, - divisorType:: hidden.core.resource.quantity, - // Required: resource to select - withResource(resource):: self + __resourceFieldRefMixin({ resource: resource }), - }, - resourceFieldRefType:: hidden.core.v1.resourceFieldSelector, - // Selects a key of a secret in the pod's namespace - secretKeyRef:: { - local __secretKeyRefMixin(secretKeyRef) = __valueFromMixin({ secretKeyRef+: secretKeyRef }), - mixinInstance(secretKeyRef):: __secretKeyRefMixin(secretKeyRef), - // The key of the secret to select from. Must be a valid secret key. - withKey(key):: self + __secretKeyRefMixin({ key: key }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretKeyRefMixin({ name: name }), - // Specify whether the Secret or it's key must be defined - withOptional(optional):: self + __secretKeyRefMixin({ optional: optional }), - }, - secretKeyRefType:: hidden.core.v1.secretKeySelector, - }, - valueFromType:: hidden.core.v1.envVarSource, - }, - }, - // EnvVarSource represents a source for the value of an EnvVar. - envVarSource:: { - new():: {}, - mixin:: { - // Selects a key of a ConfigMap. - configMapKeyRef:: { - local __configMapKeyRefMixin(configMapKeyRef) = { configMapKeyRef+: configMapKeyRef }, - mixinInstance(configMapKeyRef):: __configMapKeyRefMixin(configMapKeyRef), - // The key to select. - withKey(key):: self + __configMapKeyRefMixin({ key: key }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __configMapKeyRefMixin({ name: name }), - // Specify whether the ConfigMap or it's key must be defined - withOptional(optional):: self + __configMapKeyRefMixin({ optional: optional }), - }, - configMapKeyRefType:: hidden.core.v1.configMapKeySelector, - // Selects a field of the pod: supports metadata.name, metadata.namespace, metadata.labels, metadata.annotations, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP. - fieldRef:: { - local __fieldRefMixin(fieldRef) = { fieldRef+: fieldRef }, - mixinInstance(fieldRef):: __fieldRefMixin(fieldRef), - // Path of the field to select in the specified API version. - withFieldPath(fieldPath):: self + __fieldRefMixin({ fieldPath: fieldPath }), - }, - fieldRefType:: hidden.core.v1.objectFieldSelector, - // Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported. - resourceFieldRef:: { - local __resourceFieldRefMixin(resourceFieldRef) = { resourceFieldRef+: resourceFieldRef }, - mixinInstance(resourceFieldRef):: __resourceFieldRefMixin(resourceFieldRef), - // Container name: required for volumes, optional for env vars - withContainerName(containerName):: self + __resourceFieldRefMixin({ containerName: containerName }), - // Specifies the output format of the exposed resources, defaults to "1" - divisor:: { - local __divisorMixin(divisor) = __resourceFieldRefMixin({ divisor+: divisor }), - mixinInstance(divisor):: __divisorMixin(divisor), - }, - divisorType:: hidden.core.resource.quantity, - // Required: resource to select - withResource(resource):: self + __resourceFieldRefMixin({ resource: resource }), - }, - resourceFieldRefType:: hidden.core.v1.resourceFieldSelector, - // Selects a key of a secret in the pod's namespace - secretKeyRef:: { - local __secretKeyRefMixin(secretKeyRef) = { secretKeyRef+: secretKeyRef }, - mixinInstance(secretKeyRef):: __secretKeyRefMixin(secretKeyRef), - // The key of the secret to select from. Must be a valid secret key. - withKey(key):: self + __secretKeyRefMixin({ key: key }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretKeyRefMixin({ name: name }), - // Specify whether the Secret or it's key must be defined - withOptional(optional):: self + __secretKeyRefMixin({ optional: optional }), - }, - secretKeyRefType:: hidden.core.v1.secretKeySelector, - }, - }, - // EventList is a list of events. - eventList:: { - new(items=''):: self.withItems(items), - // List of events - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of events - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.event, - mixin:: {}, - }, - // EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time. - eventSeries:: { - new():: {}, - // Number of occurrences in this series up to the last heartbeat time - withCount(count):: self + { count: count }, - // Time of the last occurrence observed - withLastObservedTime(lastObservedTime):: self + { lastObservedTime: lastObservedTime }, - // State of this Series: Ongoing or Finished - withState(state):: self + { state: state }, - mixin:: {}, - }, - // EventSource contains information for an event. - eventSource:: { - new():: {}, - // Component from which the event is generated. - withComponent(component):: self + { component: component }, - // Node name on which the event is generated. - withHost(host):: self + { host: host }, - mixin:: {}, - }, - // ExecAction describes a "run in container" action. - execAction:: { - new():: {}, - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then { command: command } else { command: [command] }, - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then { command+: command } else { command+: [command] }, - mixin:: {}, - }, - // Represents a Fibre Channel volume. Fibre Channel volumes can only be mounted as read/write once. Fibre Channel volumes support ownership management and SELinux relabeling. - fcVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Optional: FC target lun number - withLun(lun):: self + { lun: lun }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Optional: FC target worldwide names (WWNs) - withTargetWwns(targetWwns):: self + if std.type(targetWwns) == 'array' then { targetWWNs: targetWwns } else { targetWWNs: [targetWwns] }, - // Optional: FC target worldwide names (WWNs) - withTargetWwnsMixin(targetWwns):: self + if std.type(targetWwns) == 'array' then { targetWWNs+: targetWwns } else { targetWWNs+: [targetWwns] }, - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwids(wwids):: self + if std.type(wwids) == 'array' then { wwids: wwids } else { wwids: [wwids] }, - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwidsMixin(wwids):: self + if std.type(wwids) == 'array' then { wwids+: wwids } else { wwids+: [wwids] }, - mixin:: {}, - }, - // FlexPersistentVolumeSource represents a generic persistent volume resource that is provisioned/attached using an exec based plugin. - flexPersistentVolumeSource:: { - new():: {}, - // Driver is the name of the driver to use for this volume. - withDriver(driver):: self + { driver: driver }, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script. - withFsType(fsType):: self + { fsType: fsType }, - // Optional: Extra command options if any. - withOptions(options):: self + { options: options }, - // Optional: Extra command options if any. - withOptionsMixin(options):: self + { options+: options }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: { - // Optional: SecretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin. - flexVolumeSource:: { - new():: {}, - // Driver is the name of the driver to use for this volume. - withDriver(driver):: self + { driver: driver }, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script. - withFsType(fsType):: self + { fsType: fsType }, - // Optional: Extra command options if any. - withOptions(options):: self + { options: options }, - // Optional: Extra command options if any. - withOptionsMixin(options):: self + { options+: options }, - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: { - // Optional: SecretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // Represents a Flocker volume mounted by the Flocker agent. One and only one of datasetName and datasetUUID should be set. Flocker volumes do not support ownership management or SELinux relabeling. - flockerVolumeSource:: { - new():: {}, - // Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated - withDatasetName(datasetName):: self + { datasetName: datasetName }, - // UUID of the dataset. This is unique identifier of a Flocker dataset - withDatasetUuid(datasetUuid):: self + { datasetUUID: datasetUuid }, - mixin:: {}, - }, - // Represents a Persistent Disk resource in Google Compute Engine. - // - // A GCE PD must exist before mounting to a container. The disk must also be in the same GCE project and zone as the kubelet. A GCE PD can only be mounted as read/write once or read-only many times. GCE PDs support ownership management and SELinux relabeling. - gcePersistentDiskVolumeSource:: { - new():: {}, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withFsType(fsType):: self + { fsType: fsType }, - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPartition(partition):: self + { partition: partition }, - // Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPdName(pdName):: self + { pdName: pdName }, - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // Represents a volume that is populated with the contents of a git repository. Git repo volumes do not support ownership management. Git repo volumes support SELinux relabeling. - // - // DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container. - gitRepoVolumeSource:: { - new():: {}, - // Target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name. - withDirectory(directory):: self + { directory: directory }, - // Repository URL - withRepository(repository):: self + { repository: repository }, - // Commit hash for the specified revision. - withRevision(revision):: self + { revision: revision }, - mixin:: {}, - }, - // Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling. - glusterfsPersistentVolumeSource:: { - new():: {}, - // EndpointsName is the endpoint name that details Glusterfs topology. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpoints(endpoints):: self + { endpoints: endpoints }, - // EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpointsNamespace(endpointsNamespace):: self + { endpointsNamespace: endpointsNamespace }, - // Path is the Glusterfs volume path. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withPath(path):: self + { path: path }, - // ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // Represents a Glusterfs mount that lasts the lifetime of a pod. Glusterfs volumes do not support ownership management or SELinux relabeling. - glusterfsVolumeSource:: { - new():: {}, - // EndpointsName is the endpoint name that details Glusterfs topology. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpoints(endpoints):: self + { endpoints: endpoints }, - // Path is the Glusterfs volume path. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withPath(path):: self + { path: path }, - // ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // Handler defines a specific action that should be taken - handler:: { - new():: {}, - mixin:: { - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = { exec+: exec }, - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = { httpGet+: httpGet }, - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = { tcpSocket+: tcpSocket }, - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - }, - // HostAlias holds the mapping between IP and hostnames that will be injected as an entry in the pod's hosts file. - hostAlias:: { - new():: {}, - // Hostnames for the above IP address. - withHostnames(hostnames):: self + if std.type(hostnames) == 'array' then { hostnames: hostnames } else { hostnames: [hostnames] }, - // Hostnames for the above IP address. - withHostnamesMixin(hostnames):: self + if std.type(hostnames) == 'array' then { hostnames+: hostnames } else { hostnames+: [hostnames] }, - // IP address of the host file entry. - withIp(ip):: self + { ip: ip }, - mixin:: {}, - }, - // Represents a host path mapped into a pod. Host path volumes do not support ownership management or SELinux relabeling. - hostPathVolumeSource:: { - new():: {}, - // Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withPath(path):: self + { path: path }, - // Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withType(type):: self + { type: type }, - mixin:: {}, - }, - // HTTPGetAction describes an action based on HTTP Get requests. - httpGetAction:: { - new():: {}, - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + { host: host }, - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then { httpHeaders: httpHeaders } else { httpHeaders: [httpHeaders] }, - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then { httpHeaders+: httpHeaders } else { httpHeaders+: [httpHeaders] }, - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + { path: path }, - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + { port: port }, - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + { scheme: scheme }, - mixin:: {}, - }, - // HTTPHeader describes a custom header to be used in HTTP probes - httpHeader:: { - new():: {}, - // The header field name - withName(name):: self + { name: name }, - // The header field value - withValue(value):: self + { value: value }, - mixin:: {}, - }, - // ISCSIPersistentVolumeSource represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling. - iscsiPersistentVolumeSource:: { - new():: {}, - // whether support iSCSI Discovery CHAP authentication - withChapAuthDiscovery(chapAuthDiscovery):: self + { chapAuthDiscovery: chapAuthDiscovery }, - // whether support iSCSI Session CHAP authentication - withChapAuthSession(chapAuthSession):: self + { chapAuthSession: chapAuthSession }, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi - withFsType(fsType):: self + { fsType: fsType }, - // Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection. - withInitiatorName(initiatorName):: self + { initiatorName: initiatorName }, - // Target iSCSI Qualified Name. - withIqn(iqn):: self + { iqn: iqn }, - // iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp). - withIscsiInterface(iscsiInterface):: self + { iscsiInterface: iscsiInterface }, - // iSCSI Target Lun number. - withLun(lun):: self + { lun: lun }, - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortals(portals):: self + if std.type(portals) == 'array' then { portals: portals } else { portals: [portals] }, - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortalsMixin(portals):: self + if std.type(portals) == 'array' then { portals+: portals } else { portals+: [portals] }, - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withTargetPortal(targetPortal):: self + { targetPortal: targetPortal }, - mixin:: { - // CHAP Secret for iSCSI target and initiator authentication - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // Represents an ISCSI disk. ISCSI volumes can only be mounted as read/write once. ISCSI volumes support ownership management and SELinux relabeling. - iscsiVolumeSource:: { - new():: {}, - // whether support iSCSI Discovery CHAP authentication - withChapAuthDiscovery(chapAuthDiscovery):: self + { chapAuthDiscovery: chapAuthDiscovery }, - // whether support iSCSI Session CHAP authentication - withChapAuthSession(chapAuthSession):: self + { chapAuthSession: chapAuthSession }, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi - withFsType(fsType):: self + { fsType: fsType }, - // Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection. - withInitiatorName(initiatorName):: self + { initiatorName: initiatorName }, - // Target iSCSI Qualified Name. - withIqn(iqn):: self + { iqn: iqn }, - // iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp). - withIscsiInterface(iscsiInterface):: self + { iscsiInterface: iscsiInterface }, - // iSCSI Target Lun number. - withLun(lun):: self + { lun: lun }, - // iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortals(portals):: self + if std.type(portals) == 'array' then { portals: portals } else { portals: [portals] }, - // iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortalsMixin(portals):: self + if std.type(portals) == 'array' then { portals+: portals } else { portals+: [portals] }, - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withTargetPortal(targetPortal):: self + { targetPortal: targetPortal }, - mixin:: { - // CHAP Secret for iSCSI target and initiator authentication - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // Maps a string key to a path within a volume. - keyToPath:: { - new(key='', path=''):: self.withKey(key).withPath(path), - // The key to project. - withKey(key):: self + { key: key }, - // Optional: mode bits to use on this file, must be a value between 0 and 0777. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withMode(mode):: self + { mode: mode }, - // The relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'. - withPath(path):: self + { path: path }, - mixin:: {}, - }, - // Lifecycle describes actions that the management system should take in response to container lifecycle events. For the PostStart and PreStop lifecycle handlers, management of the container blocks until the action is complete, unless the container process fails, in which case the handler is aborted. - lifecycle:: { - new():: {}, - mixin:: { - // PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks - postStart:: { - local __postStartMixin(postStart) = { postStart+: postStart }, - mixinInstance(postStart):: __postStartMixin(postStart), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __postStartMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __postStartMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __postStartMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - postStartType:: hidden.core.v1.handler, - // PreStop is called immediately before a container is terminated due to an API request or management event such as liveness probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The reason for termination is passed to the handler. The Pod's termination grace period countdown begins before the PreStop hooked is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period. Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks - preStop:: { - local __preStopMixin(preStop) = { preStop+: preStop }, - mixinInstance(preStop):: __preStopMixin(preStop), - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = __preStopMixin({ exec+: exec }), - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = __preStopMixin({ httpGet+: httpGet }), - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = __preStopMixin({ tcpSocket+: tcpSocket }), - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - preStopType:: hidden.core.v1.handler, - }, - }, - // LimitRangeItem defines a min/max usage limit for any resource that matches on kind. - limitRangeItem:: { - new():: {}, - // Default resource requirement limit value by resource name if resource limit is omitted. - withDefault(default):: self + { default: default }, - // Default resource requirement limit value by resource name if resource limit is omitted. - withDefaultMixin(default):: self + { default+: default }, - // DefaultRequest is the default resource requirement request value by resource name if resource request is omitted. - withDefaultRequest(defaultRequest):: self + { defaultRequest: defaultRequest }, - // DefaultRequest is the default resource requirement request value by resource name if resource request is omitted. - withDefaultRequestMixin(defaultRequest):: self + { defaultRequest+: defaultRequest }, - // Max usage constraints on this kind by resource name. - withMax(max):: self + { max: max }, - // Max usage constraints on this kind by resource name. - withMaxMixin(max):: self + { max+: max }, - // MaxLimitRequestRatio if specified, the named resource must have a request and limit that are both non-zero where limit divided by request is less than or equal to the enumerated value; this represents the max burst for the named resource. - withMaxLimitRequestRatio(maxLimitRequestRatio):: self + { maxLimitRequestRatio: maxLimitRequestRatio }, - // MaxLimitRequestRatio if specified, the named resource must have a request and limit that are both non-zero where limit divided by request is less than or equal to the enumerated value; this represents the max burst for the named resource. - withMaxLimitRequestRatioMixin(maxLimitRequestRatio):: self + { maxLimitRequestRatio+: maxLimitRequestRatio }, - // Min usage constraints on this kind by resource name. - withMin(min):: self + { min: min }, - // Min usage constraints on this kind by resource name. - withMinMixin(min):: self + { min+: min }, - // Type of resource that this limit applies to. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // LimitRangeList is a list of LimitRange items. - limitRangeList:: { - new(items=''):: self.withItems(items), - // Items is a list of LimitRange objects. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of LimitRange objects. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.limitRange, - mixin:: {}, - }, - // LimitRangeSpec defines a min/max usage limit for resources that match on kind. - limitRangeSpec:: { - new():: {}, - // Limits is the list of LimitRangeItem objects that are enforced. - withLimits(limits):: self + if std.type(limits) == 'array' then { limits: limits } else { limits: [limits] }, - // Limits is the list of LimitRangeItem objects that are enforced. - withLimitsMixin(limits):: self + if std.type(limits) == 'array' then { limits+: limits } else { limits+: [limits] }, - limitsType:: hidden.core.v1.limitRangeItem, - mixin:: {}, - }, - // LoadBalancerIngress represents the status of a load-balancer ingress point: traffic intended for the service should be sent to an ingress point. - loadBalancerIngress:: { - new():: {}, - // Hostname is set for load-balancer ingress points that are DNS based (typically AWS load-balancers) - withHostname(hostname):: self + { hostname: hostname }, - // IP is set for load-balancer ingress points that are IP based (typically GCE or OpenStack load-balancers) - withIp(ip):: self + { ip: ip }, - mixin:: {}, - }, - // LoadBalancerStatus represents the status of a load-balancer. - loadBalancerStatus:: { - new():: {}, - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngress(ingress):: self + if std.type(ingress) == 'array' then { ingress: ingress } else { ingress: [ingress] }, - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then { ingress+: ingress } else { ingress+: [ingress] }, - ingressType:: hidden.core.v1.loadBalancerIngress, - mixin:: {}, - }, - // LocalObjectReference contains enough information to let you locate the referenced object inside the same namespace. - localObjectReference:: { - new():: {}, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Local represents directly-attached storage with node affinity (Beta feature) - localVolumeSource:: { - new():: {}, - // Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...). - withPath(path):: self + { path: path }, - mixin:: {}, - }, - // NamespaceList is a list of Namespaces. - namespaceList:: { - new(items=''):: self.withItems(items), - // Items is the list of Namespace objects in the list. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Namespace objects in the list. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.namespace, - mixin:: {}, - }, - // NamespaceSpec describes the attributes on a Namespace. - namespaceSpec:: { - new():: {}, - // Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/ - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then { finalizers: finalizers } else { finalizers: [finalizers] }, - // Finalizers is an opaque list of values that must be empty to permanently remove object from storage. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/ - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then { finalizers+: finalizers } else { finalizers+: [finalizers] }, - mixin:: {}, - }, - // NamespaceStatus is information about the current status of a Namespace. - namespaceStatus:: { - new():: {}, - // Phase is the current lifecycle phase of the namespace. More info: https://kubernetes.io/docs/tasks/administer-cluster/namespaces/ - withPhase(phase):: self + { phase: phase }, - mixin:: {}, - }, - // Represents an NFS mount that lasts the lifetime of a pod. NFS volumes do not support ownership management or SELinux relabeling. - nfsVolumeSource:: { - new():: {}, - // Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withPath(path):: self + { path: path }, - // ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withServer(server):: self + { server: server }, - mixin:: {}, - }, - // NodeAddress contains information for the node's address. - nodeAddress:: { - new():: {}, - // The node address. - withAddress(address):: self + { address: address }, - // Node address type, one of Hostname, ExternalIP or InternalIP. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // Node affinity is a group of node affinity scheduling rules. - nodeAffinity:: { - new():: {}, - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }, - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }, - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - mixin:: { - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = { requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }, - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - }, - // NodeCondition contains condition information for a node. - nodeCondition:: { - new():: {}, - // Last time we got an update on a given condition. - withLastHeartbeatTime(lastHeartbeatTime):: self + { lastHeartbeatTime: lastHeartbeatTime }, - // Last time the condition transit from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // (brief) reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of node condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // NodeConfigSource specifies a source of node configuration. Exactly one subfield (excluding metadata) must be non-nil. - nodeConfigSource:: { - new():: {}, - mixin:: { - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = { configMap+: configMap }, - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - }, - // NodeConfigStatus describes the status of the config assigned by Node.Spec.ConfigSource. - nodeConfigStatus:: { - new():: {}, - // Error describes any problems reconciling the Spec.ConfigSource to the Active config. Errors may occur, for example, attempting to checkpoint Spec.ConfigSource to the local Assigned record, attempting to checkpoint the payload associated with Spec.ConfigSource, attempting to load or validate the Assigned config, etc. Errors may occur at different points while syncing config. Earlier errors (e.g. download or checkpointing errors) will not result in a rollback to LastKnownGood, and may resolve across Kubelet retries. Later errors (e.g. loading or validating a checkpointed config) will result in a rollback to LastKnownGood. In the latter case, it is usually possible to resolve the error by fixing the config assigned in Spec.ConfigSource. You can find additional information for debugging by searching the error message in the Kubelet log. Error is a human-readable description of the error state; machines can check whether or not Error is empty, but should not rely on the stability of the Error text across Kubelet versions. - withErrorParam(errorParam):: self + { "error": errorParam }, - mixin:: { - // Active reports the checkpointed config the node is actively using. Active will represent either the current version of the Assigned config, or the current LastKnownGood config, depending on whether attempting to use the Assigned config results in an error. - active:: { - local __activeMixin(active) = { active+: active }, - mixinInstance(active):: __activeMixin(active), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __activeMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - activeType:: hidden.core.v1.nodeConfigSource, - // Assigned reports the checkpointed config the node will try to use. When Node.Spec.ConfigSource is updated, the node checkpoints the associated config payload to local disk, along with a record indicating intended config. The node refers to this record to choose its config checkpoint, and reports this record in Assigned. Assigned only updates in the status after the record has been checkpointed to disk. When the Kubelet is restarted, it tries to make the Assigned config the Active config by loading and validating the checkpointed payload identified by Assigned. - assigned:: { - local __assignedMixin(assigned) = { assigned+: assigned }, - mixinInstance(assigned):: __assignedMixin(assigned), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __assignedMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - assignedType:: hidden.core.v1.nodeConfigSource, - // LastKnownGood reports the checkpointed config the node will fall back to when it encounters an error attempting to use the Assigned config. The Assigned config becomes the LastKnownGood config when the node determines that the Assigned config is stable and correct. This is currently implemented as a 10-minute soak period starting when the local record of Assigned config is updated. If the Assigned config is Active at the end of this period, it becomes the LastKnownGood. Note that if Spec.ConfigSource is reset to nil (use local defaults), the LastKnownGood is also immediately reset to nil, because the local default config is always assumed good. You should not make assumptions about the node's method of determining config stability and correctness, as this may change or become configurable in the future. - lastKnownGood:: { - local __lastKnownGoodMixin(lastKnownGood) = { lastKnownGood+: lastKnownGood }, - mixinInstance(lastKnownGood):: __lastKnownGoodMixin(lastKnownGood), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __lastKnownGoodMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - lastKnownGoodType:: hidden.core.v1.nodeConfigSource, - }, - }, - // NodeDaemonEndpoints lists ports opened by daemons running on the Node. - nodeDaemonEndpoints:: { - new():: {}, - mixin:: { - // Endpoint on which Kubelet is listening. - kubeletEndpoint:: { - local __kubeletEndpointMixin(kubeletEndpoint) = { kubeletEndpoint+: kubeletEndpoint }, - mixinInstance(kubeletEndpoint):: __kubeletEndpointMixin(kubeletEndpoint), - // Port number of the given endpoint. - withPort(port):: self + __kubeletEndpointMixin({ Port: port }), - }, - kubeletEndpointType:: hidden.core.v1.daemonEndpoint, - }, - }, - // NodeList is the whole list of all Nodes which have been registered with master. - nodeList:: { - new(items=''):: self.withItems(items), - // List of nodes - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of nodes - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.node, - mixin:: {}, - }, - // A node selector represents the union of the results of one or more label queries over a set of nodes; that is, it represents the OR of the selectors represented by the node selector terms. - nodeSelector:: { - new():: {}, - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then { nodeSelectorTerms: nodeSelectorTerms } else { nodeSelectorTerms: [nodeSelectorTerms] }, - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then { nodeSelectorTerms+: nodeSelectorTerms } else { nodeSelectorTerms+: [nodeSelectorTerms] }, - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - mixin:: {}, - }, - // A node selector requirement is a selector that contains values, a key, and an operator that relates the key and values. - nodeSelectorRequirement:: { - new():: {}, - // The label key that the selector applies to. - withKey(key):: self + { key: key }, - // Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt. - withOperator(operator):: self + { operator: operator }, - // An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch. - withValues(values):: self + if std.type(values) == 'array' then { values: values } else { values: [values] }, - // An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch. - withValuesMixin(values):: self + if std.type(values) == 'array' then { values+: values } else { values+: [values] }, - mixin:: {}, - }, - // A null or empty node selector term matches no objects. The requirements of them are ANDed. The TopologySelectorTerm type implements a subset of the NodeSelectorTerm. - nodeSelectorTerm:: { - new():: {}, - // A list of node selector requirements by node's labels. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions: matchExpressions } else { matchExpressions: [matchExpressions] }, - // A list of node selector requirements by node's labels. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions+: matchExpressions } else { matchExpressions+: [matchExpressions] }, - matchExpressionsType:: hidden.core.v1.nodeSelectorRequirement, - // A list of node selector requirements by node's fields. - withMatchFields(matchFields):: self + if std.type(matchFields) == 'array' then { matchFields: matchFields } else { matchFields: [matchFields] }, - // A list of node selector requirements by node's fields. - withMatchFieldsMixin(matchFields):: self + if std.type(matchFields) == 'array' then { matchFields+: matchFields } else { matchFields+: [matchFields] }, - matchFieldsType:: hidden.core.v1.nodeSelectorRequirement, - mixin:: {}, - }, - // NodeSpec describes the attributes that a node is created with. - nodeSpec:: { - new():: {}, - // Deprecated. Not all kubelets will set this field. Remove field after 1.13. see: https://issues.k8s.io/61966 - withExternalId(externalId):: self + { externalID: externalId }, - // PodCIDR represents the pod IP range assigned to the node. - withPodCidr(podCidr):: self + { podCIDR: podCidr }, - // ID of the node assigned by the cloud provider in the format: :// - withProviderId(providerId):: self + { providerID: providerId }, - // If specified, the node's taints. - withTaints(taints):: self + if std.type(taints) == 'array' then { taints: taints } else { taints: [taints] }, - // If specified, the node's taints. - withTaintsMixin(taints):: self + if std.type(taints) == 'array' then { taints+: taints } else { taints+: [taints] }, - taintsType:: hidden.core.v1.taint, - // Unschedulable controls node schedulability of new pods. By default, node is schedulable. More info: https://kubernetes.io/docs/concepts/nodes/node/#manual-node-administration - withUnschedulable(unschedulable):: self + { unschedulable: unschedulable }, - mixin:: { - // If specified, the source to get node configuration from The DynamicKubeletConfig feature gate must be enabled for the Kubelet to use this field - configSource:: { - local __configSourceMixin(configSource) = { configSource+: configSource }, - mixinInstance(configSource):: __configSourceMixin(configSource), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __configSourceMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - configSourceType:: hidden.core.v1.nodeConfigSource, - }, - }, - // NodeStatus is information about the current status of a node. - nodeStatus:: { - new():: {}, - // List of addresses reachable to the node. Queried from cloud provider, if available. More info: https://kubernetes.io/docs/concepts/nodes/node/#addresses - withAddresses(addresses):: self + if std.type(addresses) == 'array' then { addresses: addresses } else { addresses: [addresses] }, - // List of addresses reachable to the node. Queried from cloud provider, if available. More info: https://kubernetes.io/docs/concepts/nodes/node/#addresses - withAddressesMixin(addresses):: self + if std.type(addresses) == 'array' then { addresses+: addresses } else { addresses+: [addresses] }, - addressesType:: hidden.core.v1.nodeAddress, - // Allocatable represents the resources of a node that are available for scheduling. Defaults to Capacity. - withAllocatable(allocatable):: self + { allocatable: allocatable }, - // Allocatable represents the resources of a node that are available for scheduling. Defaults to Capacity. - withAllocatableMixin(allocatable):: self + { allocatable+: allocatable }, - // Capacity represents the total resources of a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacity(capacity):: self + { capacity: capacity }, - // Capacity represents the total resources of a node. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacityMixin(capacity):: self + { capacity+: capacity }, - // Conditions is an array of current observed node conditions. More info: https://kubernetes.io/docs/concepts/nodes/node/#condition - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Conditions is an array of current observed node conditions. More info: https://kubernetes.io/docs/concepts/nodes/node/#condition - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.core.v1.nodeCondition, - // List of container images on this node - withImages(images):: self + if std.type(images) == 'array' then { images: images } else { images: [images] }, - // List of container images on this node - withImagesMixin(images):: self + if std.type(images) == 'array' then { images+: images } else { images+: [images] }, - imagesType:: hidden.core.v1.containerImage, - // NodePhase is the recently observed lifecycle phase of the node. More info: https://kubernetes.io/docs/concepts/nodes/node/#phase The field is never populated, and now is deprecated. - withPhase(phase):: self + { phase: phase }, - // List of volumes that are attached to the node. - withVolumesAttached(volumesAttached):: self + if std.type(volumesAttached) == 'array' then { volumesAttached: volumesAttached } else { volumesAttached: [volumesAttached] }, - // List of volumes that are attached to the node. - withVolumesAttachedMixin(volumesAttached):: self + if std.type(volumesAttached) == 'array' then { volumesAttached+: volumesAttached } else { volumesAttached+: [volumesAttached] }, - volumesAttachedType:: hidden.core.v1.attachedVolume, - // List of attachable volumes in use (mounted) by the node. - withVolumesInUse(volumesInUse):: self + if std.type(volumesInUse) == 'array' then { volumesInUse: volumesInUse } else { volumesInUse: [volumesInUse] }, - // List of attachable volumes in use (mounted) by the node. - withVolumesInUseMixin(volumesInUse):: self + if std.type(volumesInUse) == 'array' then { volumesInUse+: volumesInUse } else { volumesInUse+: [volumesInUse] }, - mixin:: { - // Status of the config assigned to the node via the dynamic Kubelet config feature. - config:: { - local __configMixin(config) = { config+: config }, - mixinInstance(config):: __configMixin(config), - // Active reports the checkpointed config the node is actively using. Active will represent either the current version of the Assigned config, or the current LastKnownGood config, depending on whether attempting to use the Assigned config results in an error. - active:: { - local __activeMixin(active) = __configMixin({ active+: active }), - mixinInstance(active):: __activeMixin(active), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __activeMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - activeType:: hidden.core.v1.nodeConfigSource, - // Assigned reports the checkpointed config the node will try to use. When Node.Spec.ConfigSource is updated, the node checkpoints the associated config payload to local disk, along with a record indicating intended config. The node refers to this record to choose its config checkpoint, and reports this record in Assigned. Assigned only updates in the status after the record has been checkpointed to disk. When the Kubelet is restarted, it tries to make the Assigned config the Active config by loading and validating the checkpointed payload identified by Assigned. - assigned:: { - local __assignedMixin(assigned) = __configMixin({ assigned+: assigned }), - mixinInstance(assigned):: __assignedMixin(assigned), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __assignedMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - assignedType:: hidden.core.v1.nodeConfigSource, - // Error describes any problems reconciling the Spec.ConfigSource to the Active config. Errors may occur, for example, attempting to checkpoint Spec.ConfigSource to the local Assigned record, attempting to checkpoint the payload associated with Spec.ConfigSource, attempting to load or validate the Assigned config, etc. Errors may occur at different points while syncing config. Earlier errors (e.g. download or checkpointing errors) will not result in a rollback to LastKnownGood, and may resolve across Kubelet retries. Later errors (e.g. loading or validating a checkpointed config) will result in a rollback to LastKnownGood. In the latter case, it is usually possible to resolve the error by fixing the config assigned in Spec.ConfigSource. You can find additional information for debugging by searching the error message in the Kubelet log. Error is a human-readable description of the error state; machines can check whether or not Error is empty, but should not rely on the stability of the Error text across Kubelet versions. - withErrorParam(errorParam):: self + __configMixin({ "error": errorParam }), - // LastKnownGood reports the checkpointed config the node will fall back to when it encounters an error attempting to use the Assigned config. The Assigned config becomes the LastKnownGood config when the node determines that the Assigned config is stable and correct. This is currently implemented as a 10-minute soak period starting when the local record of Assigned config is updated. If the Assigned config is Active at the end of this period, it becomes the LastKnownGood. Note that if Spec.ConfigSource is reset to nil (use local defaults), the LastKnownGood is also immediately reset to nil, because the local default config is always assumed good. You should not make assumptions about the node's method of determining config stability and correctness, as this may change or become configurable in the future. - lastKnownGood:: { - local __lastKnownGoodMixin(lastKnownGood) = __configMixin({ lastKnownGood+: lastKnownGood }), - mixinInstance(lastKnownGood):: __lastKnownGoodMixin(lastKnownGood), - // ConfigMap is a reference to a Node's ConfigMap - configMap:: { - local __configMapMixin(configMap) = __lastKnownGoodMixin({ configMap+: configMap }), - mixinInstance(configMap):: __configMapMixin(configMap), - // KubeletConfigKey declares which key of the referenced ConfigMap corresponds to the KubeletConfiguration structure This field is required in all cases. - withKubeletConfigKey(kubeletConfigKey):: self + __configMapMixin({ kubeletConfigKey: kubeletConfigKey }), - // Name is the metadata.name of the referenced ConfigMap. This field is required in all cases. - withName(name):: self + __configMapMixin({ name: name }), - // Namespace is the metadata.namespace of the referenced ConfigMap. This field is required in all cases. - withNamespace(namespace):: self + __configMapMixin({ namespace: namespace }), - // ResourceVersion is the metadata.ResourceVersion of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withResourceVersion(resourceVersion):: self + __configMapMixin({ resourceVersion: resourceVersion }), - // UID is the metadata.UID of the referenced ConfigMap. This field is forbidden in Node.Spec, and required in Node.Status. - withUid(uid):: self + __configMapMixin({ uid: uid }), - }, - configMapType:: hidden.core.v1.configMapNodeConfigSource, - }, - lastKnownGoodType:: hidden.core.v1.nodeConfigSource, - }, - configType:: hidden.core.v1.nodeConfigStatus, - // Endpoints of daemons running on the Node. - daemonEndpoints:: { - local __daemonEndpointsMixin(daemonEndpoints) = { daemonEndpoints+: daemonEndpoints }, - mixinInstance(daemonEndpoints):: __daemonEndpointsMixin(daemonEndpoints), - // Endpoint on which Kubelet is listening. - kubeletEndpoint:: { - local __kubeletEndpointMixin(kubeletEndpoint) = __daemonEndpointsMixin({ kubeletEndpoint+: kubeletEndpoint }), - mixinInstance(kubeletEndpoint):: __kubeletEndpointMixin(kubeletEndpoint), - // Port number of the given endpoint. - withPort(port):: self + __kubeletEndpointMixin({ Port: port }), - }, - kubeletEndpointType:: hidden.core.v1.daemonEndpoint, - }, - daemonEndpointsType:: hidden.core.v1.nodeDaemonEndpoints, - // Set of ids/uuids to uniquely identify the node. More info: https://kubernetes.io/docs/concepts/nodes/node/#info - nodeInfo:: { - local __nodeInfoMixin(nodeInfo) = { nodeInfo+: nodeInfo }, - mixinInstance(nodeInfo):: __nodeInfoMixin(nodeInfo), - // The Architecture reported by the node - withArchitecture(architecture):: self + __nodeInfoMixin({ architecture: architecture }), - // Boot ID reported by the node. - withBootId(bootId):: self + __nodeInfoMixin({ bootID: bootId }), - // ContainerRuntime Version reported by the node through runtime remote API (e.g. docker://1.5.0). - withContainerRuntimeVersion(containerRuntimeVersion):: self + __nodeInfoMixin({ containerRuntimeVersion: containerRuntimeVersion }), - // Kernel Version reported by the node from 'uname -r' (e.g. 3.16.0-0.bpo.4-amd64). - withKernelVersion(kernelVersion):: self + __nodeInfoMixin({ kernelVersion: kernelVersion }), - // KubeProxy Version reported by the node. - withKubeProxyVersion(kubeProxyVersion):: self + __nodeInfoMixin({ kubeProxyVersion: kubeProxyVersion }), - // Kubelet Version reported by the node. - withKubeletVersion(kubeletVersion):: self + __nodeInfoMixin({ kubeletVersion: kubeletVersion }), - // MachineID reported by the node. For unique machine identification in the cluster this field is preferred. Learn more from man(5) machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html - withMachineId(machineId):: self + __nodeInfoMixin({ machineID: machineId }), - // The Operating System reported by the node - withOperatingSystem(operatingSystem):: self + __nodeInfoMixin({ operatingSystem: operatingSystem }), - // OS Image reported by the node from /etc/os-release (e.g. Debian GNU/Linux 7 (wheezy)). - withOsImage(osImage):: self + __nodeInfoMixin({ osImage: osImage }), - // SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html - withSystemUuid(systemUuid):: self + __nodeInfoMixin({ systemUUID: systemUuid }), - }, - nodeInfoType:: hidden.core.v1.nodeSystemInfo, - }, - }, - // NodeSystemInfo is a set of ids/uuids to uniquely identify the node. - nodeSystemInfo:: { - new():: {}, - // The Architecture reported by the node - withArchitecture(architecture):: self + { architecture: architecture }, - // Boot ID reported by the node. - withBootId(bootId):: self + { bootID: bootId }, - // ContainerRuntime Version reported by the node through runtime remote API (e.g. docker://1.5.0). - withContainerRuntimeVersion(containerRuntimeVersion):: self + { containerRuntimeVersion: containerRuntimeVersion }, - // Kernel Version reported by the node from 'uname -r' (e.g. 3.16.0-0.bpo.4-amd64). - withKernelVersion(kernelVersion):: self + { kernelVersion: kernelVersion }, - // KubeProxy Version reported by the node. - withKubeProxyVersion(kubeProxyVersion):: self + { kubeProxyVersion: kubeProxyVersion }, - // Kubelet Version reported by the node. - withKubeletVersion(kubeletVersion):: self + { kubeletVersion: kubeletVersion }, - // MachineID reported by the node. For unique machine identification in the cluster this field is preferred. Learn more from man(5) machine-id: http://man7.org/linux/man-pages/man5/machine-id.5.html - withMachineId(machineId):: self + { machineID: machineId }, - // The Operating System reported by the node - withOperatingSystem(operatingSystem):: self + { operatingSystem: operatingSystem }, - // OS Image reported by the node from /etc/os-release (e.g. Debian GNU/Linux 7 (wheezy)). - withOsImage(osImage):: self + { osImage: osImage }, - // SystemUUID reported by the node. For unique machine identification MachineID is preferred. This field is specific to Red Hat hosts https://access.redhat.com/documentation/en-US/Red_Hat_Subscription_Management/1/html/RHSM/getting-system-uuid.html - withSystemUuid(systemUuid):: self + { systemUUID: systemUuid }, - mixin:: {}, - }, - // ObjectFieldSelector selects an APIVersioned field of an object. - objectFieldSelector:: { - new():: {}, - // Path of the field to select in the specified API version. - withFieldPath(fieldPath):: self + { fieldPath: fieldPath }, - mixin:: {}, - }, - // ObjectReference contains enough information to let you inspect or modify the referred object. - objectReference:: { - new():: {}, - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + { fieldPath: fieldPath }, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + { namespace: namespace }, - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + { resourceVersion: resourceVersion }, - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + { uid: uid }, - mixin:: {}, - }, - // PersistentVolumeClaimCondition contails details about state of pvc - persistentVolumeClaimCondition:: { - new():: {}, - // Last time we probed the condition. - withLastProbeTime(lastProbeTime):: self + { lastProbeTime: lastProbeTime }, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human-readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // Unique, this should be a short, machine understandable string that gives the reason for condition's last transition. If it reports "ResizeStarted" that means the underlying persistent volume is being resized. - withReason(reason):: self + { reason: reason }, - withStatus(status):: self + { status: status }, - withType(type):: self + { type: type }, - mixin:: {}, - }, - // PersistentVolumeClaimList is a list of PersistentVolumeClaim items. - persistentVolumeClaimList:: { - new(items=''):: self.withItems(items), - // A list of persistent volume claims. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // A list of persistent volume claims. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.persistentVolumeClaim, - mixin:: {}, - }, - // PersistentVolumeClaimSpec describes the common attributes of storage devices and allows a Source for provider-specific attributes - persistentVolumeClaimSpec:: { - new():: {}, - // AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModes(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes: accessModes } else { accessModes: [accessModes] }, - // AccessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModesMixin(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes+: accessModes } else { accessModes+: [accessModes] }, - // Name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1 - withStorageClassName(storageClassName):: self + { storageClassName: storageClassName }, - // volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec. This is a beta feature. - withVolumeMode(volumeMode):: self + { volumeMode: volumeMode }, - // VolumeName is the binding reference to the PersistentVolume backing this claim. - withVolumeName(volumeName):: self + { volumeName: volumeName }, - mixin:: { - // This field requires the VolumeSnapshotDataSource alpha feature gate to be enabled and currently VolumeSnapshot is the only supported data source. If the provisioner can support VolumeSnapshot data source, it will create a new volume and data will be restored to the volume at the same time. If the provisioner does not support VolumeSnapshot data source, volume will not be created and the failure will be reported as an event. In the future, we plan to support more data source types and the behavior of the provisioner may change. - dataSource:: { - local __dataSourceMixin(dataSource) = { dataSource+: dataSource }, - mixinInstance(dataSource):: __dataSourceMixin(dataSource), - // APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required. - withApiGroup(apiGroup):: self + __dataSourceMixin({ apiGroup: apiGroup }), - // Kind is the type of resource being referenced - withKind(kind):: self + __dataSourceMixin({ kind: kind }), - // Name is the name of resource being referenced - withName(name):: self + __dataSourceMixin({ name: name }), - }, - dataSourceType:: hidden.core.v1.typedLocalObjectReference, - // Resources represents the minimum resources the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources - resources:: { - local __resourcesMixin(resources) = { resources+: resources }, - mixinInstance(resources):: __resourcesMixin(resources), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimits(limits):: self + __resourcesMixin({ limits: limits }), - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimitsMixin(limits):: self + __resourcesMixin({ limits+: limits }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequests(requests):: self + __resourcesMixin({ requests: requests }), - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequestsMixin(requests):: self + __resourcesMixin({ requests+: requests }), - }, - resourcesType:: hidden.core.v1.resourceRequirements, - // A label query over volumes to consider for binding. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - }, - // PersistentVolumeClaimStatus is the current status of a persistent volume claim. - persistentVolumeClaimStatus:: { - new():: {}, - // AccessModes contains the actual access modes the volume backing the PVC has. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModes(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes: accessModes } else { accessModes: [accessModes] }, - // AccessModes contains the actual access modes the volume backing the PVC has. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1 - withAccessModesMixin(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes+: accessModes } else { accessModes+: [accessModes] }, - // Represents the actual resources of the underlying volume. - withCapacity(capacity):: self + { capacity: capacity }, - // Represents the actual resources of the underlying volume. - withCapacityMixin(capacity):: self + { capacity+: capacity }, - // Current Condition of persistent volume claim. If underlying persistent volume is being resized then the Condition will be set to 'ResizeStarted'. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Current Condition of persistent volume claim. If underlying persistent volume is being resized then the Condition will be set to 'ResizeStarted'. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.core.v1.persistentVolumeClaimCondition, - // Phase represents the current phase of PersistentVolumeClaim. - withPhase(phase):: self + { phase: phase }, - mixin:: {}, - }, - // PersistentVolumeClaimVolumeSource references the user's PVC in the same namespace. This volume finds the bound PV and mounts that volume for the pod. A PersistentVolumeClaimVolumeSource is, essentially, a wrapper around another type of volume that is owned by someone else (the system). - persistentVolumeClaimVolumeSource:: { - new():: {}, - // ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - withClaimName(claimName):: self + { claimName: claimName }, - // Will force the ReadOnly setting in VolumeMounts. Default false. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // PersistentVolumeList is a list of PersistentVolume items. - persistentVolumeList:: { - new(items=''):: self.withItems(items), - // List of persistent volumes. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of persistent volumes. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.persistentVolume, - mixin:: {}, - }, - // PersistentVolumeSpec is the specification of a persistent volume. - persistentVolumeSpec:: { - new():: {}, - // AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes - withAccessModes(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes: accessModes } else { accessModes: [accessModes] }, - // AccessModes contains all ways the volume can be mounted. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes - withAccessModesMixin(accessModes):: self + if std.type(accessModes) == 'array' then { accessModes+: accessModes } else { accessModes+: [accessModes] }, - // A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacity(capacity):: self + { capacity: capacity }, - // A description of the persistent volume's resources and capacity. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#capacity - withCapacityMixin(capacity):: self + { capacity+: capacity }, - // A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options - withMountOptions(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions: mountOptions } else { mountOptions: [mountOptions] }, - // A list of mount options, e.g. ["ro", "soft"]. Not validated - mount will simply fail if one is invalid. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#mount-options - withMountOptionsMixin(mountOptions):: self + if std.type(mountOptions) == 'array' then { mountOptions+: mountOptions } else { mountOptions+: [mountOptions] }, - // What happens to a persistent volume when released from its claim. Valid options are Retain (default for manually created PersistentVolumes), Delete (default for dynamically provisioned PersistentVolumes), and Recycle (deprecated). Recycle must be supported by the volume plugin underlying this PersistentVolume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#reclaiming - withPersistentVolumeReclaimPolicy(persistentVolumeReclaimPolicy):: self + { persistentVolumeReclaimPolicy: persistentVolumeReclaimPolicy }, - // Name of StorageClass to which this persistent volume belongs. Empty value means that this volume does not belong to any StorageClass. - withStorageClassName(storageClassName):: self + { storageClassName: storageClassName }, - // volumeMode defines if a volume is intended to be used with a formatted filesystem or to remain in raw block state. Value of Filesystem is implied when not included in spec. This is a beta feature. - withVolumeMode(volumeMode):: self + { volumeMode: volumeMode }, - mixin:: { - // AWSElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - awsElasticBlockStore:: { - local __awsElasticBlockStoreMixin(awsElasticBlockStore) = { awsElasticBlockStore+: awsElasticBlockStore }, - mixinInstance(awsElasticBlockStore):: __awsElasticBlockStoreMixin(awsElasticBlockStore), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withFsType(fsType):: self + __awsElasticBlockStoreMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). - withPartition(partition):: self + __awsElasticBlockStoreMixin({ partition: partition }), - // Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withReadOnly(readOnly):: self + __awsElasticBlockStoreMixin({ readOnly: readOnly }), - // Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withVolumeId(volumeId):: self + __awsElasticBlockStoreMixin({ volumeID: volumeId }), - }, - awsElasticBlockStoreType:: hidden.core.v1.awsElasticBlockStoreVolumeSource, - // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod. - azureDisk:: { - local __azureDiskMixin(azureDisk) = { azureDisk+: azureDisk }, - mixinInstance(azureDisk):: __azureDiskMixin(azureDisk), - // Host Caching mode: None, Read Only, Read Write. - withCachingMode(cachingMode):: self + __azureDiskMixin({ cachingMode: cachingMode }), - // The Name of the data disk in the blob storage - withDiskName(diskName):: self + __azureDiskMixin({ diskName: diskName }), - // The URI the data disk in the blob storage - withDiskUri(diskUri):: self + __azureDiskMixin({ diskURI: diskUri }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __azureDiskMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureDiskMixin({ readOnly: readOnly }), - }, - azureDiskType:: hidden.core.v1.azureDiskVolumeSource, - // AzureFile represents an Azure File Service mount on the host and bind mount to the pod. - azureFile:: { - local __azureFileMixin(azureFile) = { azureFile+: azureFile }, - mixinInstance(azureFile):: __azureFileMixin(azureFile), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureFileMixin({ readOnly: readOnly }), - // the name of secret that contains Azure Storage Account Name and Key - withSecretName(secretName):: self + __azureFileMixin({ secretName: secretName }), - // the namespace of the secret that contains Azure Storage Account Name and Key default is the same as the Pod - withSecretNamespace(secretNamespace):: self + __azureFileMixin({ secretNamespace: secretNamespace }), - // Share Name - withShareName(shareName):: self + __azureFileMixin({ shareName: shareName }), - }, - azureFileType:: hidden.core.v1.azureFilePersistentVolumeSource, - // CephFS represents a Ceph FS mount on the host that shares a pod's lifetime - cephfs:: { - local __cephfsMixin(cephfs) = { cephfs+: cephfs }, - mixinInstance(cephfs):: __cephfsMixin(cephfs), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors: monitors }) else __cephfsMixin({ monitors: [monitors] }), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors+: monitors }) else __cephfsMixin({ monitors+: [monitors] }), - // Optional: Used as the mounted root, rather than the full Ceph tree, default is / - withPath(path):: self + __cephfsMixin({ path: path }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withReadOnly(readOnly):: self + __cephfsMixin({ readOnly: readOnly }), - // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withSecretFile(secretFile):: self + __cephfsMixin({ secretFile: secretFile }), - // Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __cephfsMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // Optional: User is the rados user name, default is admin More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withUser(user):: self + __cephfsMixin({ user: user }), - }, - cephfsType:: hidden.core.v1.cephFsPersistentVolumeSource, - // Cinder represents a cinder volume attached and mounted on kubelets host machine More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - cinder:: { - local __cinderMixin(cinder) = { cinder+: cinder }, - mixinInstance(cinder):: __cinderMixin(cinder), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withFsType(fsType):: self + __cinderMixin({ fsType: fsType }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withReadOnly(readOnly):: self + __cinderMixin({ readOnly: readOnly }), - // Optional: points to a secret object containing parameters used to connect to OpenStack. - secretRef:: { - local __secretRefMixin(secretRef) = __cinderMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // volume id used to identify the volume in cinder More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withVolumeId(volumeId):: self + __cinderMixin({ volumeID: volumeId }), - }, - cinderType:: hidden.core.v1.cinderPersistentVolumeSource, - // ClaimRef is part of a bi-directional binding between PersistentVolume and PersistentVolumeClaim. Expected to be non-nil when bound. claim.VolumeName is the authoritative bind between PV and PVC. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#binding - claimRef:: { - local __claimRefMixin(claimRef) = { claimRef+: claimRef }, - mixinInstance(claimRef):: __claimRefMixin(claimRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __claimRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __claimRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __claimRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __claimRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __claimRefMixin({ uid: uid }), - }, - claimRefType:: hidden.core.v1.objectReference, - // CSI represents storage that is handled by an external CSI driver (Beta feature). - csi:: { - local __csiMixin(csi) = { csi+: csi }, - mixinInstance(csi):: __csiMixin(csi), - // ControllerPublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI ControllerPublishVolume and ControllerUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - controllerPublishSecretRef:: { - local __controllerPublishSecretRefMixin(controllerPublishSecretRef) = __csiMixin({ controllerPublishSecretRef+: controllerPublishSecretRef }), - mixinInstance(controllerPublishSecretRef):: __controllerPublishSecretRefMixin(controllerPublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __controllerPublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __controllerPublishSecretRefMixin({ namespace: namespace }), - }, - controllerPublishSecretRefType:: hidden.core.v1.secretReference, - // Driver is the name of the driver to use for this volume. Required. - withDriver(driver):: self + __csiMixin({ driver: driver }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". - withFsType(fsType):: self + __csiMixin({ fsType: fsType }), - // NodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodePublishSecretRef:: { - local __nodePublishSecretRefMixin(nodePublishSecretRef) = __csiMixin({ nodePublishSecretRef+: nodePublishSecretRef }), - mixinInstance(nodePublishSecretRef):: __nodePublishSecretRefMixin(nodePublishSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodePublishSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodePublishSecretRefMixin({ namespace: namespace }), - }, - nodePublishSecretRefType:: hidden.core.v1.secretReference, - // NodeStageSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodeStageVolume and NodeStageVolume and NodeUnstageVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secrets are passed. - nodeStageSecretRef:: { - local __nodeStageSecretRefMixin(nodeStageSecretRef) = __csiMixin({ nodeStageSecretRef+: nodeStageSecretRef }), - mixinInstance(nodeStageSecretRef):: __nodeStageSecretRefMixin(nodeStageSecretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __nodeStageSecretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __nodeStageSecretRefMixin({ namespace: namespace }), - }, - nodeStageSecretRefType:: hidden.core.v1.secretReference, - // Optional: The value to pass to ControllerPublishVolumeRequest. Defaults to false (read/write). - withReadOnly(readOnly):: self + __csiMixin({ readOnly: readOnly }), - // Attributes of the volume to publish. - withVolumeAttributes(volumeAttributes):: self + __csiMixin({ volumeAttributes: volumeAttributes }), - // Attributes of the volume to publish. - withVolumeAttributesMixin(volumeAttributes):: self + __csiMixin({ volumeAttributes+: volumeAttributes }), - // VolumeHandle is the unique volume name returned by the CSI volume plugin’s CreateVolume to refer to the volume on all subsequent calls. Required. - withVolumeHandle(volumeHandle):: self + __csiMixin({ volumeHandle: volumeHandle }), - }, - csiType:: hidden.core.v1.csiPersistentVolumeSource, - // FC represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod. - fc:: { - local __fcMixin(fc) = { fc+: fc }, - mixinInstance(fc):: __fcMixin(fc), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __fcMixin({ fsType: fsType }), - // Optional: FC target lun number - withLun(lun):: self + __fcMixin({ lun: lun }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __fcMixin({ readOnly: readOnly }), - // Optional: FC target worldwide names (WWNs) - withTargetWwns(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs: targetWwns }) else __fcMixin({ targetWWNs: [targetWwns] }), - // Optional: FC target worldwide names (WWNs) - withTargetWwnsMixin(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs+: targetWwns }) else __fcMixin({ targetWWNs+: [targetWwns] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwids(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids: wwids }) else __fcMixin({ wwids: [wwids] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwidsMixin(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids+: wwids }) else __fcMixin({ wwids+: [wwids] }), - }, - fcType:: hidden.core.v1.fcVolumeSource, - // FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin. - flexVolume:: { - local __flexVolumeMixin(flexVolume) = { flexVolume+: flexVolume }, - mixinInstance(flexVolume):: __flexVolumeMixin(flexVolume), - // Driver is the name of the driver to use for this volume. - withDriver(driver):: self + __flexVolumeMixin({ driver: driver }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script. - withFsType(fsType):: self + __flexVolumeMixin({ fsType: fsType }), - // Optional: Extra command options if any. - withOptions(options):: self + __flexVolumeMixin({ options: options }), - // Optional: Extra command options if any. - withOptionsMixin(options):: self + __flexVolumeMixin({ options+: options }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __flexVolumeMixin({ readOnly: readOnly }), - // Optional: SecretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts. - secretRef:: { - local __secretRefMixin(secretRef) = __flexVolumeMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - flexVolumeType:: hidden.core.v1.flexPersistentVolumeSource, - // Flocker represents a Flocker volume attached to a kubelet's host machine and exposed to the pod for its usage. This depends on the Flocker control service being running - flocker:: { - local __flockerMixin(flocker) = { flocker+: flocker }, - mixinInstance(flocker):: __flockerMixin(flocker), - // Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated - withDatasetName(datasetName):: self + __flockerMixin({ datasetName: datasetName }), - // UUID of the dataset. This is unique identifier of a Flocker dataset - withDatasetUuid(datasetUuid):: self + __flockerMixin({ datasetUUID: datasetUuid }), - }, - flockerType:: hidden.core.v1.flockerVolumeSource, - // GCEPersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. Provisioned by an admin. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - gcePersistentDisk:: { - local __gcePersistentDiskMixin(gcePersistentDisk) = { gcePersistentDisk+: gcePersistentDisk }, - mixinInstance(gcePersistentDisk):: __gcePersistentDiskMixin(gcePersistentDisk), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withFsType(fsType):: self + __gcePersistentDiskMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPartition(partition):: self + __gcePersistentDiskMixin({ partition: partition }), - // Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPdName(pdName):: self + __gcePersistentDiskMixin({ pdName: pdName }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withReadOnly(readOnly):: self + __gcePersistentDiskMixin({ readOnly: readOnly }), - }, - gcePersistentDiskType:: hidden.core.v1.gcePersistentDiskVolumeSource, - // Glusterfs represents a Glusterfs volume that is attached to a host and exposed to the pod. Provisioned by an admin. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md - glusterfs:: { - local __glusterfsMixin(glusterfs) = { glusterfs+: glusterfs }, - mixinInstance(glusterfs):: __glusterfsMixin(glusterfs), - // EndpointsName is the endpoint name that details Glusterfs topology. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpoints(endpoints):: self + __glusterfsMixin({ endpoints: endpoints }), - // EndpointsNamespace is the namespace that contains Glusterfs endpoint. If this field is empty, the EndpointNamespace defaults to the same namespace as the bound PVC. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpointsNamespace(endpointsNamespace):: self + __glusterfsMixin({ endpointsNamespace: endpointsNamespace }), - // Path is the Glusterfs volume path. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withPath(path):: self + __glusterfsMixin({ path: path }), - // ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withReadOnly(readOnly):: self + __glusterfsMixin({ readOnly: readOnly }), - }, - glusterfsType:: hidden.core.v1.glusterfsPersistentVolumeSource, - // HostPath represents a directory on the host. Provisioned by a developer or tester. This is useful for single-node development and testing only! On-host storage is not supported in any way and WILL NOT WORK in a multi-node cluster. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - hostPath:: { - local __hostPathMixin(hostPath) = { hostPath+: hostPath }, - mixinInstance(hostPath):: __hostPathMixin(hostPath), - // Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withPath(path):: self + __hostPathMixin({ path: path }), - // Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withType(type):: self + __hostPathMixin({ type: type }), - }, - hostPathType:: hidden.core.v1.hostPathVolumeSource, - // ISCSI represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. Provisioned by an admin. - iscsi:: { - local __iscsiMixin(iscsi) = { iscsi+: iscsi }, - mixinInstance(iscsi):: __iscsiMixin(iscsi), - // whether support iSCSI Discovery CHAP authentication - withChapAuthDiscovery(chapAuthDiscovery):: self + __iscsiMixin({ chapAuthDiscovery: chapAuthDiscovery }), - // whether support iSCSI Session CHAP authentication - withChapAuthSession(chapAuthSession):: self + __iscsiMixin({ chapAuthSession: chapAuthSession }), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi - withFsType(fsType):: self + __iscsiMixin({ fsType: fsType }), - // Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection. - withInitiatorName(initiatorName):: self + __iscsiMixin({ initiatorName: initiatorName }), - // Target iSCSI Qualified Name. - withIqn(iqn):: self + __iscsiMixin({ iqn: iqn }), - // iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp). - withIscsiInterface(iscsiInterface):: self + __iscsiMixin({ iscsiInterface: iscsiInterface }), - // iSCSI Target Lun number. - withLun(lun):: self + __iscsiMixin({ lun: lun }), - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortals(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals: portals }) else __iscsiMixin({ portals: [portals] }), - // iSCSI Target Portal List. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortalsMixin(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals+: portals }) else __iscsiMixin({ portals+: [portals] }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. - withReadOnly(readOnly):: self + __iscsiMixin({ readOnly: readOnly }), - // CHAP Secret for iSCSI target and initiator authentication - secretRef:: { - local __secretRefMixin(secretRef) = __iscsiMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withTargetPortal(targetPortal):: self + __iscsiMixin({ targetPortal: targetPortal }), - }, - iscsiType:: hidden.core.v1.iscsiPersistentVolumeSource, - // Local represents directly-attached storage with node affinity - localStorage:: { - local __localStorageMixin(localStorage) = { localStorage+: localStorage }, - mixinInstance(localStorage):: __localStorageMixin(localStorage), - // Filesystem type to mount. It applies only when the Path is a block device. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default value is to auto-select a fileystem if unspecified. - withFsType(fsType):: self + __localStorageMixin({ fsType: fsType }), - // The full path to the volume on the node. It can be either a directory or block device (disk, partition, ...). - withPath(path):: self + __localStorageMixin({ path: path }), - }, - localType:: hidden.core.v1.localVolumeSource, - // NFS represents an NFS mount on the host. Provisioned by an admin. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - nfs:: { - local __nfsMixin(nfs) = { nfs+: nfs }, - mixinInstance(nfs):: __nfsMixin(nfs), - // Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withPath(path):: self + __nfsMixin({ path: path }), - // ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withReadOnly(readOnly):: self + __nfsMixin({ readOnly: readOnly }), - // Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withServer(server):: self + __nfsMixin({ server: server }), - }, - nfsType:: hidden.core.v1.nfsVolumeSource, - // NodeAffinity defines constraints that limit what nodes this volume can be accessed from. This field influences the scheduling of pods that use this volume. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = { nodeAffinity+: nodeAffinity }, - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // Required specifies hard node constraints that must be met. - required:: { - local __requiredMixin(required) = __nodeAffinityMixin({ required+: required }), - mixinInstance(required):: __requiredMixin(required), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.volumeNodeAffinity, - // PhotonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine - photonPersistentDisk:: { - local __photonPersistentDiskMixin(photonPersistentDisk) = { photonPersistentDisk+: photonPersistentDisk }, - mixinInstance(photonPersistentDisk):: __photonPersistentDiskMixin(photonPersistentDisk), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __photonPersistentDiskMixin({ fsType: fsType }), - // ID that identifies Photon Controller persistent disk - withPdId(pdId):: self + __photonPersistentDiskMixin({ pdID: pdId }), - }, - photonPersistentDiskType:: hidden.core.v1.photonPersistentDiskVolumeSource, - // PortworxVolume represents a portworx volume attached and mounted on kubelets host machine - portworxVolume:: { - local __portworxVolumeMixin(portworxVolume) = { portworxVolume+: portworxVolume }, - mixinInstance(portworxVolume):: __portworxVolumeMixin(portworxVolume), - // FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __portworxVolumeMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __portworxVolumeMixin({ readOnly: readOnly }), - // VolumeID uniquely identifies a Portworx volume - withVolumeId(volumeId):: self + __portworxVolumeMixin({ volumeID: volumeId }), - }, - portworxVolumeType:: hidden.core.v1.portworxVolumeSource, - // Quobyte represents a Quobyte mount on the host that shares a pod's lifetime - quobyte:: { - local __quobyteMixin(quobyte) = { quobyte+: quobyte }, - mixinInstance(quobyte):: __quobyteMixin(quobyte), - // Group to map volume access to Default is no group - withGroup(group):: self + __quobyteMixin({ group: group }), - // ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false. - withReadOnly(readOnly):: self + __quobyteMixin({ readOnly: readOnly }), - // Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes - withRegistry(registry):: self + __quobyteMixin({ registry: registry }), - // Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin - withTenant(tenant):: self + __quobyteMixin({ tenant: tenant }), - // User to map volume access to Defaults to serivceaccount user - withUser(user):: self + __quobyteMixin({ user: user }), - // Volume is a string that references an already created Quobyte volume by name. - withVolume(volume):: self + __quobyteMixin({ volume: volume }), - }, - quobyteType:: hidden.core.v1.quobyteVolumeSource, - // RBD represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md - rbd:: { - local __rbdMixin(rbd) = { rbd+: rbd }, - mixinInstance(rbd):: __rbdMixin(rbd), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd - withFsType(fsType):: self + __rbdMixin({ fsType: fsType }), - // The rados image name. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withImage(image):: self + __rbdMixin({ image: image }), - // Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withKeyring(keyring):: self + __rbdMixin({ keyring: keyring }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors: monitors }) else __rbdMixin({ monitors: [monitors] }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors+: monitors }) else __rbdMixin({ monitors+: [monitors] }), - // The rados pool name. Default is rbd. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withPool(pool):: self + __rbdMixin({ pool: pool }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withReadOnly(readOnly):: self + __rbdMixin({ readOnly: readOnly }), - // SecretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __rbdMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // The rados user name. Default is admin. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withUser(user):: self + __rbdMixin({ user: user }), - }, - rbdType:: hidden.core.v1.rbdPersistentVolumeSource, - // ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes. - scaleIo:: { - local __scaleIoMixin(scaleIo) = { scaleIo+: scaleIo }, - mixinInstance(scaleIo):: __scaleIoMixin(scaleIo), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs" - withFsType(fsType):: self + __scaleIoMixin({ fsType: fsType }), - // The host address of the ScaleIO API Gateway. - withGateway(gateway):: self + __scaleIoMixin({ gateway: gateway }), - // The name of the ScaleIO Protection Domain for the configured storage. - withProtectionDomain(protectionDomain):: self + __scaleIoMixin({ protectionDomain: protectionDomain }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __scaleIoMixin({ readOnly: readOnly }), - // SecretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail. - secretRef:: { - local __secretRefMixin(secretRef) = __scaleIoMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - // Flag to enable/disable SSL communication with Gateway, default false - withSslEnabled(sslEnabled):: self + __scaleIoMixin({ sslEnabled: sslEnabled }), - // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned. - withStorageMode(storageMode):: self + __scaleIoMixin({ storageMode: storageMode }), - // The ScaleIO Storage Pool associated with the protection domain. - withStoragePool(storagePool):: self + __scaleIoMixin({ storagePool: storagePool }), - // The name of the storage system as configured in ScaleIO. - withSystem(system):: self + __scaleIoMixin({ system: system }), - // The name of a volume already created in the ScaleIO system that is associated with this volume source. - withVolumeName(volumeName):: self + __scaleIoMixin({ volumeName: volumeName }), - }, - scaleIOType:: hidden.core.v1.scaleIoPersistentVolumeSource, - // StorageOS represents a StorageOS volume that is attached to the kubelet's host machine and mounted into the pod More info: https://releases.k8s.io/HEAD/examples/volumes/storageos/README.md - storageos:: { - local __storageosMixin(storageos) = { storageos+: storageos }, - mixinInstance(storageos):: __storageosMixin(storageos), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __storageosMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __storageosMixin({ readOnly: readOnly }), - // SecretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted. - secretRef:: { - local __secretRefMixin(secretRef) = __storageosMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __secretRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __secretRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __secretRefMixin({ uid: uid }), - }, - secretRefType:: hidden.core.v1.objectReference, - // VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace. - withVolumeName(volumeName):: self + __storageosMixin({ volumeName: volumeName }), - // VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to "default" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created. - withVolumeNamespace(volumeNamespace):: self + __storageosMixin({ volumeNamespace: volumeNamespace }), - }, - storageosType:: hidden.core.v1.storageOsPersistentVolumeSource, - // VsphereVolume represents a vSphere volume attached and mounted on kubelets host machine - vsphereVolume:: { - local __vsphereVolumeMixin(vsphereVolume) = { vsphereVolume+: vsphereVolume }, - mixinInstance(vsphereVolume):: __vsphereVolumeMixin(vsphereVolume), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __vsphereVolumeMixin({ fsType: fsType }), - // Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName. - withStoragePolicyId(storagePolicyId):: self + __vsphereVolumeMixin({ storagePolicyID: storagePolicyId }), - // Storage Policy Based Management (SPBM) profile name. - withStoragePolicyName(storagePolicyName):: self + __vsphereVolumeMixin({ storagePolicyName: storagePolicyName }), - // Path that identifies vSphere volume vmdk - withVolumePath(volumePath):: self + __vsphereVolumeMixin({ volumePath: volumePath }), - }, - vsphereVolumeType:: hidden.core.v1.vsphereVirtualDiskVolumeSource, - }, - }, - // PersistentVolumeStatus is the current status of a persistent volume. - persistentVolumeStatus:: { - new():: {}, - // A human-readable message indicating details about why the volume is in this state. - withMessage(message):: self + { message: message }, - // Phase indicates if a volume is available, bound to a claim, or released by a claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#phase - withPhase(phase):: self + { phase: phase }, - // Reason is a brief CamelCase string that describes any failure and is meant for machine parsing and tidy display in the CLI. - withReason(reason):: self + { reason: reason }, - mixin:: {}, - }, - // Represents a Photon Controller persistent disk resource. - photonPersistentDiskVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // ID that identifies Photon Controller persistent disk - withPdId(pdId):: self + { pdID: pdId }, - mixin:: {}, - }, - // Pod affinity is a group of inter pod affinity scheduling rules. - podAffinity:: { - new():: {}, - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }, - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }, - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then { requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution } else { requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then { requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution } else { requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - mixin:: {}, - }, - // Defines a set of pods (namely those matching the labelSelector relative to the given namespace(s)) that this pod should be co-located (affinity) or not co-located (anti-affinity) with, where co-located is defined as running on a node whose value of the label with key matches that of any node on which a pod of the set of pods is running - podAffinityTerm:: { - new():: {}, - // namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means "this pod's namespace" - withNamespaces(namespaces):: self + if std.type(namespaces) == 'array' then { namespaces: namespaces } else { namespaces: [namespaces] }, - // namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means "this pod's namespace" - withNamespacesMixin(namespaces):: self + if std.type(namespaces) == 'array' then { namespaces+: namespaces } else { namespaces+: [namespaces] }, - // This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed. - withTopologyKey(topologyKey):: self + { topologyKey: topologyKey }, - mixin:: { - // A label query over a set of resources, in this case pods. - labelSelector:: { - local __labelSelectorMixin(labelSelector) = { labelSelector+: labelSelector }, - mixinInstance(labelSelector):: __labelSelectorMixin(labelSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __labelSelectorMixin({ matchExpressions: matchExpressions }) else __labelSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __labelSelectorMixin({ matchExpressions+: matchExpressions }) else __labelSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __labelSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __labelSelectorMixin({ matchLabels+: matchLabels }), - }, - labelSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // Pod anti affinity is a group of inter pod anti affinity scheduling rules. - podAntiAffinity:: { - new():: {}, - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }, - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then { preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution } else { preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }, - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then { requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution } else { requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then { requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution } else { requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - mixin:: {}, - }, - // PodCondition contains details for the current condition of this pod. - podCondition:: { - new():: {}, - // Last time we probed the condition. - withLastProbeTime(lastProbeTime):: self + { lastProbeTime: lastProbeTime }, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // Human-readable message indicating details about last transition. - withMessage(message):: self + { message: message }, - // Unique, one-word, CamelCase reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status is the status of the condition. Can be True, False, Unknown. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions - withStatus(status):: self + { status: status }, - // Type is the type of the condition. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions - withType(type):: self + { type: type }, - mixin:: {}, - }, - // PodDNSConfig defines the DNS parameters of a pod in addition to those generated from DNSPolicy. - podDnsConfig:: { - new():: {}, - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then { nameservers: nameservers } else { nameservers: [nameservers] }, - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then { nameservers+: nameservers } else { nameservers+: [nameservers] }, - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then { options: options } else { options: [options] }, - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then { options+: options } else { options+: [options] }, - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then { searches: searches } else { searches: [searches] }, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then { searches+: searches } else { searches+: [searches] }, - mixin:: {}, - }, - // PodDNSConfigOption defines DNS resolver options of a pod. - podDnsConfigOption:: { - new():: {}, - // Required. - withName(name):: self + { name: name }, - withValue(value):: self + { value: value }, - mixin:: {}, - }, - // PodList is a list of Pods. - podList:: { - new(items=''):: self.withItems(items), - // List of pods. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of pods. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.pod, - mixin:: {}, - }, - // PodReadinessGate contains the reference to a pod condition - podReadinessGate:: { - new():: {}, - // ConditionType refers to a condition in the pod's condition list with matching type. - withConditionType(conditionType):: self + { conditionType: conditionType }, - mixin:: {}, - }, - // PodSecurityContext holds pod-level security attributes and common container settings. Some fields are also present in container.securityContext. Field values of container.securityContext take precedence over field values of PodSecurityContext. - podSecurityContext:: { - new():: {}, - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + { fsGroup: fsGroup }, - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + { runAsGroup: runAsGroup }, - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + { runAsNonRoot: runAsNonRoot }, - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + { runAsUser: runAsUser }, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then { supplementalGroups: supplementalGroups } else { supplementalGroups: [supplementalGroups] }, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then { supplementalGroups+: supplementalGroups } else { supplementalGroups+: [supplementalGroups] }, - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then { sysctls: sysctls } else { sysctls: [sysctls] }, - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then { sysctls+: sysctls } else { sysctls+: [sysctls] }, - sysctlsType:: hidden.core.v1.sysctl, - mixin:: { - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = { seLinuxOptions+: seLinuxOptions }, - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - }, - // PodSpec is a description of a pod. - podSpec:: { - new():: {}, - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + { activeDeadlineSeconds: activeDeadlineSeconds }, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + { automountServiceAccountToken: automountServiceAccountToken }, - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then { containers: containers } else { containers: [containers] }, - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then { containers+: containers } else { containers+: [containers] }, - containersType:: hidden.core.v1.container, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + { dnsPolicy: dnsPolicy }, - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + { enableServiceLinks: enableServiceLinks }, - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then { hostAliases: hostAliases } else { hostAliases: [hostAliases] }, - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then { hostAliases+: hostAliases } else { hostAliases+: [hostAliases] }, - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + { hostIPC: hostIpc }, - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + { hostNetwork: hostNetwork }, - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + { hostPID: hostPid }, - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + { hostname: hostname }, - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then { imagePullSecrets: imagePullSecrets } else { imagePullSecrets: [imagePullSecrets] }, - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then { imagePullSecrets+: imagePullSecrets } else { imagePullSecrets+: [imagePullSecrets] }, - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then { initContainers: initContainers } else { initContainers: [initContainers] }, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then { initContainers+: initContainers } else { initContainers+: [initContainers] }, - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + { nodeName: nodeName }, - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + { nodeSelector: nodeSelector }, - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + { nodeSelector+: nodeSelector }, - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + { priority: priority }, - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + { priorityClassName: priorityClassName }, - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then { readinessGates: readinessGates } else { readinessGates: [readinessGates] }, - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then { readinessGates+: readinessGates } else { readinessGates+: [readinessGates] }, - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + { restartPolicy: restartPolicy }, - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + { runtimeClassName: runtimeClassName }, - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + { schedulerName: schedulerName }, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + { serviceAccount: serviceAccount }, - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + { serviceAccountName: serviceAccountName }, - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + { shareProcessNamespace: shareProcessNamespace }, - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + { subdomain: subdomain }, - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + { terminationGracePeriodSeconds: terminationGracePeriodSeconds }, - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then { tolerations: tolerations } else { tolerations: [tolerations] }, - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then { tolerations+: tolerations } else { tolerations+: [tolerations] }, - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then { volumes: volumes } else { volumes: [volumes] }, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then { volumes+: volumes } else { volumes+: [volumes] }, - volumesType:: hidden.core.v1.volume, - mixin:: { - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = { affinity+: affinity }, - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = { dnsConfig+: dnsConfig }, - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = { securityContext+: securityContext }, - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - }, - }, - // PodStatus represents information about the status of a pod. Status may trail the actual state of a system, especially if the node that hosts the pod cannot contact the control plane. - podStatus:: { - new():: {}, - // Current service state of pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Current service state of pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-conditions - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.core.v1.podCondition, - // The list has one entry per container in the manifest. Each entry is currently the output of `docker inspect`. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status - withContainerStatuses(containerStatuses):: self + if std.type(containerStatuses) == 'array' then { containerStatuses: containerStatuses } else { containerStatuses: [containerStatuses] }, - // The list has one entry per container in the manifest. Each entry is currently the output of `docker inspect`. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status - withContainerStatusesMixin(containerStatuses):: self + if std.type(containerStatuses) == 'array' then { containerStatuses+: containerStatuses } else { containerStatuses+: [containerStatuses] }, - containerStatusesType:: hidden.core.v1.containerStatus, - // IP address of the host to which the pod is assigned. Empty if not yet scheduled. - withHostIp(hostIp):: self + { hostIP: hostIp }, - // The list has one entry per init container in the manifest. The most recent successful init container will have ready = true, the most recently started container will have startTime set. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status - withInitContainerStatuses(initContainerStatuses):: self + if std.type(initContainerStatuses) == 'array' then { initContainerStatuses: initContainerStatuses } else { initContainerStatuses: [initContainerStatuses] }, - // The list has one entry per init container in the manifest. The most recent successful init container will have ready = true, the most recently started container will have startTime set. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-and-container-status - withInitContainerStatusesMixin(initContainerStatuses):: self + if std.type(initContainerStatuses) == 'array' then { initContainerStatuses+: initContainerStatuses } else { initContainerStatuses+: [initContainerStatuses] }, - initContainerStatusesType:: hidden.core.v1.containerStatus, - // A human readable message indicating details about why the pod is in this condition. - withMessage(message):: self + { message: message }, - // nominatedNodeName is set only when this pod preempts other pods on the node, but it cannot be scheduled right away as preemption victims receive their graceful termination periods. This field does not guarantee that the pod will be scheduled on this node. Scheduler may decide to place the pod elsewhere if other nodes become available sooner. Scheduler may also decide to give the resources on this node to a higher priority pod that is created after preemption. As a result, this field may be different than PodSpec.nodeName when the pod is scheduled. - withNominatedNodeName(nominatedNodeName):: self + { nominatedNodeName: nominatedNodeName }, - // The phase of a Pod is a simple, high-level summary of where the Pod is in its lifecycle. The conditions array, the reason and message fields, and the individual container status arrays contain more detail about the pod's status. There are five possible phase values: - // - // Pending: The pod has been accepted by the Kubernetes system, but one or more of the container images has not been created. This includes time before being scheduled as well as time spent downloading images over the network, which could take a while. Running: The pod has been bound to a node, and all of the containers have been created. At least one container is still running, or is in the process of starting or restarting. Succeeded: All containers in the pod have terminated in success, and will not be restarted. Failed: All containers in the pod have terminated, and at least one container has terminated in failure. The container either exited with non-zero status or was terminated by the system. Unknown: For some reason the state of the pod could not be obtained, typically due to an error in communicating with the host of the pod. - // - // More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#pod-phase - withPhase(phase):: self + { phase: phase }, - // IP address allocated to the pod. Routable at least within the cluster. Empty if not yet allocated. - withPodIp(podIp):: self + { podIP: podIp }, - // The Quality of Service (QOS) classification assigned to the pod based on resource requirements See PodQOSClass type for available QOS classes More info: https://git.k8s.io/community/contributors/design-proposals/node/resource-qos.md - withQosClass(qosClass):: self + { qosClass: qosClass }, - // A brief CamelCase message indicating details about why the pod is in this state. e.g. 'Evicted' - withReason(reason):: self + { reason: reason }, - // RFC 3339 date and time at which the object was acknowledged by the Kubelet. This is before the Kubelet pulled the container image(s) for the pod. - withStartTime(startTime):: self + { startTime: startTime }, - mixin:: {}, - }, - // PodTemplateList is a list of PodTemplates. - podTemplateList:: { - new(items=''):: self.withItems(items), - // List of pod templates - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of pod templates - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.podTemplate, - mixin:: {}, - }, - // PodTemplateSpec describes the data a pod should have when created from a template - podTemplateSpec:: { - new():: {}, - mixin:: { - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = { metadata+: metadata }, - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = { spec+: spec }, - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - }, - // PortworxVolumeSource represents a Portworx volume resource. - portworxVolumeSource:: { - new():: {}, - // FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // VolumeID uniquely identifies a Portworx volume - withVolumeId(volumeId):: self + { volumeID: volumeId }, - mixin:: {}, - }, - // An empty preferred scheduling term matches all objects with implicit weight 0 (i.e. it's a no-op). A null preferred scheduling term matches no objects (i.e. is also a no-op). - preferredSchedulingTerm:: { - new():: {}, - // Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100. - withWeight(weight):: self + { weight: weight }, - mixin:: { - // A node selector term, associated with the corresponding weight. - preference:: { - local __preferenceMixin(preference) = { preference+: preference }, - mixinInstance(preference):: __preferenceMixin(preference), - // A list of node selector requirements by node's labels. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __preferenceMixin({ matchExpressions: matchExpressions }) else __preferenceMixin({ matchExpressions: [matchExpressions] }), - // A list of node selector requirements by node's labels. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __preferenceMixin({ matchExpressions+: matchExpressions }) else __preferenceMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.core.v1.nodeSelectorRequirement, - // A list of node selector requirements by node's fields. - withMatchFields(matchFields):: self + if std.type(matchFields) == 'array' then __preferenceMixin({ matchFields: matchFields }) else __preferenceMixin({ matchFields: [matchFields] }), - // A list of node selector requirements by node's fields. - withMatchFieldsMixin(matchFields):: self + if std.type(matchFields) == 'array' then __preferenceMixin({ matchFields+: matchFields }) else __preferenceMixin({ matchFields+: [matchFields] }), - matchFieldsType:: hidden.core.v1.nodeSelectorRequirement, - }, - preferenceType:: hidden.core.v1.nodeSelectorTerm, - }, - }, - // Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. - probe:: { - new():: {}, - // Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1. - withFailureThreshold(failureThreshold):: self + { failureThreshold: failureThreshold }, - // Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withInitialDelaySeconds(initialDelaySeconds):: self + { initialDelaySeconds: initialDelaySeconds }, - // How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1. - withPeriodSeconds(periodSeconds):: self + { periodSeconds: periodSeconds }, - // Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness. Minimum value is 1. - withSuccessThreshold(successThreshold):: self + { successThreshold: successThreshold }, - // Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes - withTimeoutSeconds(timeoutSeconds):: self + { timeoutSeconds: timeoutSeconds }, - mixin:: { - // One and only one of the following should be specified. Exec specifies the action to take. - exec:: { - local __execMixin(exec) = { exec+: exec }, - mixinInstance(exec):: __execMixin(exec), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommand(command):: self + if std.type(command) == 'array' then __execMixin({ command: command }) else __execMixin({ command: [command] }), - // Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. - withCommandMixin(command):: self + if std.type(command) == 'array' then __execMixin({ command+: command }) else __execMixin({ command+: [command] }), - }, - execType:: hidden.core.v1.execAction, - // HTTPGet specifies the http request to perform. - httpGet:: { - local __httpGetMixin(httpGet) = { httpGet+: httpGet }, - mixinInstance(httpGet):: __httpGetMixin(httpGet), - // Host name to connect to, defaults to the pod IP. You probably want to set "Host" in httpHeaders instead. - withHost(host):: self + __httpGetMixin({ host: host }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeaders(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders: httpHeaders }) else __httpGetMixin({ httpHeaders: [httpHeaders] }), - // Custom headers to set in the request. HTTP allows repeated headers. - withHttpHeadersMixin(httpHeaders):: self + if std.type(httpHeaders) == 'array' then __httpGetMixin({ httpHeaders+: httpHeaders }) else __httpGetMixin({ httpHeaders+: [httpHeaders] }), - httpHeadersType:: hidden.core.v1.httpHeader, - // Path to access on the HTTP server. - withPath(path):: self + __httpGetMixin({ path: path }), - // Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __httpGetMixin({ port: port }), - // Scheme to use for connecting to the host. Defaults to HTTP. - withScheme(scheme):: self + __httpGetMixin({ scheme: scheme }), - }, - httpGetType:: hidden.core.v1.httpGetAction, - // TCPSocket specifies an action involving a TCP port. TCP hooks not yet supported - tcpSocket:: { - local __tcpSocketMixin(tcpSocket) = { tcpSocket+: tcpSocket }, - mixinInstance(tcpSocket):: __tcpSocketMixin(tcpSocket), - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + __tcpSocketMixin({ host: host }), - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + __tcpSocketMixin({ port: port }), - }, - tcpSocketType:: hidden.core.v1.tcpSocketAction, - }, - }, - // Represents a projected volume source - projectedVolumeSource:: { - new():: {}, - // Mode bits to use on created files by default. Must be a value between 0 and 0777. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + { defaultMode: defaultMode }, - // list of volume projections - withSources(sources):: self + if std.type(sources) == 'array' then { sources: sources } else { sources: [sources] }, - // list of volume projections - withSourcesMixin(sources):: self + if std.type(sources) == 'array' then { sources+: sources } else { sources+: [sources] }, - sourcesType:: hidden.core.v1.volumeProjection, - mixin:: {}, - }, - // Represents a Quobyte mount that lasts the lifetime of a pod. Quobyte volumes do not support ownership management or SELinux relabeling. - quobyteVolumeSource:: { - new():: {}, - // Group to map volume access to Default is no group - withGroup(group):: self + { group: group }, - // ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes - withRegistry(registry):: self + { registry: registry }, - // Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin - withTenant(tenant):: self + { tenant: tenant }, - // User to map volume access to Defaults to serivceaccount user - withUser(user):: self + { user: user }, - // Volume is a string that references an already created Quobyte volume by name. - withVolume(volume):: self + { volume: volume }, - mixin:: {}, - }, - // Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling. - rbdPersistentVolumeSource:: { - new():: {}, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd - withFsType(fsType):: self + { fsType: fsType }, - // The rados image name. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withImage(image):: self + { image: image }, - // Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withKeyring(keyring):: self + { keyring: keyring }, - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then { monitors: monitors } else { monitors: [monitors] }, - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then { monitors+: monitors } else { monitors+: [monitors] }, - // The rados pool name. Default is rbd. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withPool(pool):: self + { pool: pool }, - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // The rados user name. Default is admin. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withUser(user):: self + { user: user }, - mixin:: { - // SecretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // Represents a Rados Block Device mount that lasts the lifetime of a pod. RBD volumes support ownership management and SELinux relabeling. - rbdVolumeSource:: { - new():: {}, - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd - withFsType(fsType):: self + { fsType: fsType }, - // The rados image name. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withImage(image):: self + { image: image }, - // Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withKeyring(keyring):: self + { keyring: keyring }, - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then { monitors: monitors } else { monitors: [monitors] }, - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then { monitors+: monitors } else { monitors+: [monitors] }, - // The rados pool name. Default is rbd. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withPool(pool):: self + { pool: pool }, - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // The rados user name. Default is admin. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withUser(user):: self + { user: user }, - mixin:: { - // SecretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // ReplicationControllerCondition describes the state of a replication controller at a certain point. - replicationControllerCondition:: { - new():: {}, - // The last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of replication controller condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // ReplicationControllerList is a collection of replication controllers. - replicationControllerList:: { - new(items=''):: self.withItems(items), - // List of replication controllers. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of replication controllers. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.replicationController, - mixin:: {}, - }, - // ReplicationControllerSpec is the specification of a replication controller. - replicationControllerSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - // Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withSelector(selector):: self + { selector: selector }, - // Selector is a label query over pods that should match the Replicas count. If Selector is empty, it is defaulted to the labels present on the Pod template. Label keys and values that must match in order to be controlled by this replication controller, if empty defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withSelectorMixin(selector):: self + { selector+: selector }, - mixin:: { - // Template is the object that describes the pod that will be created if insufficient replicas are detected. This takes precedence over a TemplateRef. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // ReplicationControllerStatus represents the current status of a replication controller. - replicationControllerStatus:: { - new():: {}, - // The number of available replicas (ready for at least minReadySeconds) for this replication controller. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Represents the latest available observations of a replication controller's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a replication controller's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.core.v1.replicationControllerCondition, - // The number of pods that have labels matching the labels of the pod template of the replication controller. - withFullyLabeledReplicas(fullyLabeledReplicas):: self + { fullyLabeledReplicas: fullyLabeledReplicas }, - // ObservedGeneration reflects the generation of the most recently observed replication controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The number of ready replicas for this replication controller. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // ResourceFieldSelector represents container resources (cpu, memory) and their output format - resourceFieldSelector:: { - new():: {}, - // Container name: required for volumes, optional for env vars - withContainerName(containerName):: self + { containerName: containerName }, - // Required: resource to select - withResource(resource):: self + { resource: resource }, - mixin:: { - // Specifies the output format of the exposed resources, defaults to "1" - divisor:: { - local __divisorMixin(divisor) = { divisor+: divisor }, - mixinInstance(divisor):: __divisorMixin(divisor), - }, - divisorType:: hidden.core.resource.quantity, - }, - }, - // ResourceQuotaList is a list of ResourceQuota items. - resourceQuotaList:: { - new(items=''):: self.withItems(items), - // Items is a list of ResourceQuota objects. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of ResourceQuota objects. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.resourceQuota, - mixin:: {}, - }, - // ResourceQuotaSpec defines the desired hard limits to enforce for Quota. - resourceQuotaSpec:: { - new():: {}, - // hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHard(hard):: self + { hard: hard }, - // hard is the set of desired hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHardMixin(hard):: self + { hard+: hard }, - // A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects. - withScopes(scopes):: self + if std.type(scopes) == 'array' then { scopes: scopes } else { scopes: [scopes] }, - // A collection of filters that must match each object tracked by a quota. If not specified, the quota matches all objects. - withScopesMixin(scopes):: self + if std.type(scopes) == 'array' then { scopes+: scopes } else { scopes+: [scopes] }, - mixin:: { - // scopeSelector is also a collection of filters like scopes that must match each object tracked by a quota but expressed using ScopeSelectorOperator in combination with possible values. For a resource to match, both scopes AND scopeSelector (if specified in spec), must be matched. - scopeSelector:: { - local __scopeSelectorMixin(scopeSelector) = { scopeSelector+: scopeSelector }, - mixinInstance(scopeSelector):: __scopeSelectorMixin(scopeSelector), - // A list of scope selector requirements by scope of the resources. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __scopeSelectorMixin({ matchExpressions: matchExpressions }) else __scopeSelectorMixin({ matchExpressions: [matchExpressions] }), - // A list of scope selector requirements by scope of the resources. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __scopeSelectorMixin({ matchExpressions+: matchExpressions }) else __scopeSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.core.v1.scopedResourceSelectorRequirement, - }, - scopeSelectorType:: hidden.core.v1.scopeSelector, - }, - }, - // ResourceQuotaStatus defines the enforced hard limits and observed use. - resourceQuotaStatus:: { - new():: {}, - // Hard is the set of enforced hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHard(hard):: self + { hard: hard }, - // Hard is the set of enforced hard limits for each named resource. More info: https://kubernetes.io/docs/concepts/policy/resource-quotas/ - withHardMixin(hard):: self + { hard+: hard }, - // Used is the current observed total usage of the resource in the namespace. - withUsed(used):: self + { used: used }, - // Used is the current observed total usage of the resource in the namespace. - withUsedMixin(used):: self + { used+: used }, - mixin:: {}, - }, - // ResourceRequirements describes the compute resource requirements. - resourceRequirements:: { - new():: {}, - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimits(limits):: self + { limits: limits }, - // Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withLimitsMixin(limits):: self + { limits+: limits }, - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequests(requests):: self + { requests: requests }, - // Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. More info: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/ - withRequestsMixin(requests):: self + { requests+: requests }, - mixin:: {}, - }, - // ScaleIOPersistentVolumeSource represents a persistent ScaleIO volume - scaleIoPersistentVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs" - withFsType(fsType):: self + { fsType: fsType }, - // The host address of the ScaleIO API Gateway. - withGateway(gateway):: self + { gateway: gateway }, - // The name of the ScaleIO Protection Domain for the configured storage. - withProtectionDomain(protectionDomain):: self + { protectionDomain: protectionDomain }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Flag to enable/disable SSL communication with Gateway, default false - withSslEnabled(sslEnabled):: self + { sslEnabled: sslEnabled }, - // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned. - withStorageMode(storageMode):: self + { storageMode: storageMode }, - // The ScaleIO Storage Pool associated with the protection domain. - withStoragePool(storagePool):: self + { storagePool: storagePool }, - // The name of the storage system as configured in ScaleIO. - withSystem(system):: self + { system: system }, - // The name of a volume already created in the ScaleIO system that is associated with this volume source. - withVolumeName(volumeName):: self + { volumeName: volumeName }, - mixin:: { - // SecretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - }, - secretRefType:: hidden.core.v1.secretReference, - }, - }, - // ScaleIOVolumeSource represents a persistent ScaleIO volume - scaleIoVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs". - withFsType(fsType):: self + { fsType: fsType }, - // The host address of the ScaleIO API Gateway. - withGateway(gateway):: self + { gateway: gateway }, - // The name of the ScaleIO Protection Domain for the configured storage. - withProtectionDomain(protectionDomain):: self + { protectionDomain: protectionDomain }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Flag to enable/disable SSL communication with Gateway, default false - withSslEnabled(sslEnabled):: self + { sslEnabled: sslEnabled }, - // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned. - withStorageMode(storageMode):: self + { storageMode: storageMode }, - // The ScaleIO Storage Pool associated with the protection domain. - withStoragePool(storagePool):: self + { storagePool: storagePool }, - // The name of the storage system as configured in ScaleIO. - withSystem(system):: self + { system: system }, - // The name of a volume already created in the ScaleIO system that is associated with this volume source. - withVolumeName(volumeName):: self + { volumeName: volumeName }, - mixin:: { - // SecretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // A scope selector represents the AND of the selectors represented by the scoped-resource selector requirements. - scopeSelector:: { - new():: {}, - // A list of scope selector requirements by scope of the resources. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions: matchExpressions } else { matchExpressions: [matchExpressions] }, - // A list of scope selector requirements by scope of the resources. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions+: matchExpressions } else { matchExpressions+: [matchExpressions] }, - matchExpressionsType:: hidden.core.v1.scopedResourceSelectorRequirement, - mixin:: {}, - }, - // A scoped-resource selector requirement is a selector that contains values, a scope name, and an operator that relates the scope name and values. - scopedResourceSelectorRequirement:: { - new():: {}, - // Represents a scope's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. - withOperator(operator):: self + { operator: operator }, - // The name of the scope that the selector applies to. - withScopeName(scopeName):: self + { scopeName: scopeName }, - // An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. - withValues(values):: self + if std.type(values) == 'array' then { values: values } else { values: [values] }, - // An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. - withValuesMixin(values):: self + if std.type(values) == 'array' then { values+: values } else { values+: [values] }, - mixin:: {}, - }, - // SELinuxOptions are the labels to be applied to the container - seLinuxOptions:: { - new():: {}, - // Level is SELinux level label that applies to the container. - withLevel(level):: self + { level: level }, - // Role is a SELinux role label that applies to the container. - withRole(role):: self + { role: role }, - // Type is a SELinux type label that applies to the container. - withType(type):: self + { type: type }, - // User is a SELinux user label that applies to the container. - withUser(user):: self + { user: user }, - mixin:: {}, - }, - // SecretEnvSource selects a Secret to populate the environment variables with. - // - // The contents of the target Secret's Data field will represent the key-value pairs as environment variables. - secretEnvSource:: { - new():: {}, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the Secret must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // SecretKeySelector selects a key of a Secret. - secretKeySelector:: { - new():: {}, - // The key of the secret to select from. Must be a valid secret key. - withKey(key):: self + { key: key }, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the Secret or it's key must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // SecretList is a list of Secret. - secretList:: { - new(items=''):: self.withItems(items), - // Items is a list of secret objects. More info: https://kubernetes.io/docs/concepts/configuration/secret - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of secret objects. More info: https://kubernetes.io/docs/concepts/configuration/secret - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.secret, - mixin:: {}, - }, - // Adapts a secret into a projected volume. - // - // The contents of the target Secret's Data field will be presented in a projected volume as files using the keys in the Data field as the file names. Note that this is identical to a secret volume source without the default mode. - secretProjection:: { - new():: {}, - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - // Specify whether the Secret or its key must be defined - withOptional(optional):: self + { optional: optional }, - mixin:: {}, - }, - // SecretReference represents a Secret Reference. It has enough information to retrieve secret in any namespace - secretReference:: { - new():: {}, - // Name is unique within a namespace to reference a secret resource. - withName(name):: self + { name: name }, - // Namespace defines the space within which the secret name must be unique. - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - // Adapts a Secret into a volume. - // - // The contents of the target Secret's Data field will be presented in a volume as files using the keys in the Data field as the file names. Secret volumes support ownership management and SELinux relabeling. - secretVolumeSource:: { - new():: {}, - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + { defaultMode: defaultMode }, - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.keyToPath, - // Specify whether the Secret or it's keys must be defined - withOptional(optional):: self + { optional: optional }, - // Name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret - withSecretName(secretName):: self + { secretName: secretName }, - mixin:: {}, - }, - // SecurityContext holds security configuration that will be applied to a container. Some fields are present in both SecurityContext and PodSecurityContext. When both are set, the values in SecurityContext take precedence. - securityContext:: { - new():: {}, - // AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + { allowPrivilegeEscalation: allowPrivilegeEscalation }, - // Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. - withPrivileged(privileged):: self + { privileged: privileged }, - // procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. - withProcMount(procMount):: self + { procMount: procMount }, - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsGroup(runAsGroup):: self + { runAsGroup: runAsGroup }, - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + { runAsNonRoot: runAsNonRoot }, - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsUser(runAsUser):: self + { runAsUser: runAsUser }, - mixin:: { - // The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. - capabilities:: { - local __capabilitiesMixin(capabilities) = { capabilities+: capabilities }, - mixinInstance(capabilities):: __capabilitiesMixin(capabilities), - // Added capabilities - withAdd(add):: self + if std.type(add) == 'array' then __capabilitiesMixin({ add: add }) else __capabilitiesMixin({ add: [add] }), - // Added capabilities - withAddMixin(add):: self + if std.type(add) == 'array' then __capabilitiesMixin({ add+: add }) else __capabilitiesMixin({ add+: [add] }), - // Removed capabilities - withDrop(drop):: self + if std.type(drop) == 'array' then __capabilitiesMixin({ drop: drop }) else __capabilitiesMixin({ drop: [drop] }), - // Removed capabilities - withDropMixin(drop):: self + if std.type(drop) == 'array' then __capabilitiesMixin({ drop+: drop }) else __capabilitiesMixin({ drop+: [drop] }), - }, - capabilitiesType:: hidden.core.v1.capabilities, - // The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = { seLinuxOptions+: seLinuxOptions }, - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - }, - // ServiceAccountList is a list of ServiceAccount objects - serviceAccountList:: { - new(items=''):: self.withItems(items), - // List of ServiceAccounts. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ServiceAccounts. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.serviceAccount, - mixin:: {}, - }, - // ServiceAccountTokenProjection represents a projected service account token volume. This projection can be used to insert a service account token into the pods runtime filesystem for use against APIs (Kubernetes API Server or otherwise). - serviceAccountTokenProjection:: { - new():: {}, - // Audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver. - withAudience(audience):: self + { audience: audience }, - // ExpirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes. - withExpirationSeconds(expirationSeconds):: self + { expirationSeconds: expirationSeconds }, - // Path is the path relative to the mount point of the file to project the token into. - withPath(path):: self + { path: path }, - mixin:: {}, - }, - // ServiceList holds a list of services. - serviceList:: { - new(items=''):: self.withItems(items), - // List of services - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of services - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.core.v1.service, - mixin:: {}, - }, - // ServicePort contains information on service's port. - servicePort:: { - new(port='', targetPort=''):: self.withPort(port).withTargetPort(targetPort), - newNamed(name='', port='', targetPort=''):: self.withName(name).withPort(port).withTargetPort(targetPort), - // The name of this port within the service. This must be a DNS_LABEL. All ports within a ServiceSpec must have unique names. This maps to the 'Name' field in EndpointPort objects. Optional if only one ServicePort is defined on this service. - withName(name):: self + { name: name }, - // The port on each node on which this service is exposed when type=NodePort or LoadBalancer. Usually assigned by the system. If specified, it will be allocated to the service if unused or else creation of the service will fail. Default is to auto-allocate a port if the ServiceType of this Service requires one. More info: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport - withNodePort(nodePort):: self + { nodePort: nodePort }, - // The port that will be exposed by this service. - withPort(port):: self + { port: port }, - // The IP protocol for this port. Supports "TCP", "UDP", and "SCTP". Default is TCP. - withProtocol(protocol):: self + { protocol: protocol }, - // Number or name of the port to access on the pods targeted by the service. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. If this is a string, it will be looked up as a named port in the target Pod's container ports. If this is not specified, the value of the 'port' field is used (an identity map). This field is ignored for services with clusterIP=None, and should be omitted or set equal to the 'port' field. More info: https://kubernetes.io/docs/concepts/services-networking/service/#defining-a-service - withTargetPort(targetPort):: self + { targetPort: targetPort }, - mixin:: {}, - }, - // ServiceSpec describes the attributes that a user creates on a service. - serviceSpec:: { - new():: {}, - // clusterIP is the IP address of the service and is usually assigned randomly by the master. If an address is specified manually and is not in use by others, it will be allocated to the service; otherwise, creation of the service will fail. This field can not be changed through updates. Valid values are "None", empty string (""), or a valid IP address. "None" can be specified for headless services when proxying is not required. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withClusterIp(clusterIp):: self + { clusterIP: clusterIp }, - // externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. - withExternalIps(externalIps):: self + if std.type(externalIps) == 'array' then { externalIPs: externalIps } else { externalIPs: [externalIps] }, - // externalIPs is a list of IP addresses for which nodes in the cluster will also accept traffic for this service. These IPs are not managed by Kubernetes. The user is responsible for ensuring that traffic arrives at a node with this IP. A common example is external load-balancers that are not part of the Kubernetes system. - withExternalIpsMixin(externalIps):: self + if std.type(externalIps) == 'array' then { externalIPs+: externalIps } else { externalIPs+: [externalIps] }, - // externalName is the external reference that kubedns or equivalent will return as a CNAME record for this service. No proxying will be involved. Must be a valid RFC-1123 hostname (https://tools.ietf.org/html/rfc1123) and requires Type to be ExternalName. - withExternalName(externalName):: self + { externalName: externalName }, - // externalTrafficPolicy denotes if this Service desires to route external traffic to node-local or cluster-wide endpoints. "Local" preserves the client source IP and avoids a second hop for LoadBalancer and Nodeport type services, but risks potentially imbalanced traffic spreading. "Cluster" obscures the client source IP and may cause a second hop to another node, but should have good overall load-spreading. - withExternalTrafficPolicy(externalTrafficPolicy):: self + { externalTrafficPolicy: externalTrafficPolicy }, - // healthCheckNodePort specifies the healthcheck nodePort for the service. If not specified, HealthCheckNodePort is created by the service api backend with the allocated nodePort. Will use user-specified nodePort value if specified by the client. Only effects when Type is set to LoadBalancer and ExternalTrafficPolicy is set to Local. - withHealthCheckNodePort(healthCheckNodePort):: self + { healthCheckNodePort: healthCheckNodePort }, - // Only applies to Service Type: LoadBalancer LoadBalancer will get created with the IP specified in this field. This feature depends on whether the underlying cloud-provider supports specifying the loadBalancerIP when a load balancer is created. This field will be ignored if the cloud-provider does not support the feature. - withLoadBalancerIp(loadBalancerIp):: self + { loadBalancerIP: loadBalancerIp }, - // If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/ - withLoadBalancerSourceRanges(loadBalancerSourceRanges):: self + if std.type(loadBalancerSourceRanges) == 'array' then { loadBalancerSourceRanges: loadBalancerSourceRanges } else { loadBalancerSourceRanges: [loadBalancerSourceRanges] }, - // If specified and supported by the platform, this will restrict traffic through the cloud-provider load-balancer will be restricted to the specified client IPs. This field will be ignored if the cloud-provider does not support the feature." More info: https://kubernetes.io/docs/tasks/access-application-cluster/configure-cloud-provider-firewall/ - withLoadBalancerSourceRangesMixin(loadBalancerSourceRanges):: self + if std.type(loadBalancerSourceRanges) == 'array' then { loadBalancerSourceRanges+: loadBalancerSourceRanges } else { loadBalancerSourceRanges+: [loadBalancerSourceRanges] }, - // The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // The list of ports that are exposed by this service. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.core.v1.servicePort, - // publishNotReadyAddresses, when set to true, indicates that DNS implementations must publish the notReadyAddresses of subsets for the Endpoints associated with the Service. The default value is false. The primary use case for setting this field is to use a StatefulSet's Headless Service to propagate SRV records for its Pods without respect to their readiness for purpose of peer discovery. - withPublishNotReadyAddresses(publishNotReadyAddresses):: self + { publishNotReadyAddresses: publishNotReadyAddresses }, - // Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/ - withSelector(selector):: self + { selector: selector }, - // Route service traffic to pods with label keys and values matching this selector. If empty or not present, the service is assumed to have an external process managing its endpoints, which Kubernetes will not modify. Only applies to types ClusterIP, NodePort, and LoadBalancer. Ignored if type is ExternalName. More info: https://kubernetes.io/docs/concepts/services-networking/service/ - withSelectorMixin(selector):: self + { selector+: selector }, - // Supports "ClientIP" and "None". Used to maintain session affinity. Enable client IP based session affinity. Must be ClientIP or None. Defaults to None. More info: https://kubernetes.io/docs/concepts/services-networking/service/#virtual-ips-and-service-proxies - withSessionAffinity(sessionAffinity):: self + { sessionAffinity: sessionAffinity }, - // type determines how the Service is exposed. Defaults to ClusterIP. Valid options are ExternalName, ClusterIP, NodePort, and LoadBalancer. "ExternalName" maps to the specified externalName. "ClusterIP" allocates a cluster-internal IP address for load-balancing to endpoints. Endpoints are determined by the selector or if that is not specified, by manual construction of an Endpoints object. If clusterIP is "None", no virtual IP is allocated and the endpoints are published as a set of endpoints rather than a stable IP. "NodePort" builds on ClusterIP and allocates a port on every node which routes to the clusterIP. "LoadBalancer" builds on NodePort and creates an external load-balancer (if supported in the current cloud) which routes to the clusterIP. More info: https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types - withType(type):: self + { type: type }, - mixin:: { - // sessionAffinityConfig contains the configurations of session affinity. - sessionAffinityConfig:: { - local __sessionAffinityConfigMixin(sessionAffinityConfig) = { sessionAffinityConfig+: sessionAffinityConfig }, - mixinInstance(sessionAffinityConfig):: __sessionAffinityConfigMixin(sessionAffinityConfig), - // clientIP contains the configurations of Client IP based session affinity. - clientIp:: { - local __clientIpMixin(clientIp) = __sessionAffinityConfigMixin({ clientIp+: clientIp }), - mixinInstance(clientIp):: __clientIpMixin(clientIp), - // timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours). - withTimeoutSeconds(timeoutSeconds):: self + __clientIpMixin({ timeoutSeconds: timeoutSeconds }), - }, - clientIPType:: hidden.core.v1.clientIpConfig, - }, - sessionAffinityConfigType:: hidden.core.v1.sessionAffinityConfig, - }, - }, - // ServiceStatus represents the current status of a service. - serviceStatus:: { - new():: {}, - mixin:: { - // LoadBalancer contains the current status of the load-balancer, if one is present. - loadBalancer:: { - local __loadBalancerMixin(loadBalancer) = { loadBalancer+: loadBalancer }, - mixinInstance(loadBalancer):: __loadBalancerMixin(loadBalancer), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngress(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress: ingress }) else __loadBalancerMixin({ ingress: [ingress] }), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress+: ingress }) else __loadBalancerMixin({ ingress+: [ingress] }), - ingressType:: hidden.core.v1.loadBalancerIngress, - }, - loadBalancerType:: hidden.core.v1.loadBalancerStatus, - }, - }, - // SessionAffinityConfig represents the configurations of session affinity. - sessionAffinityConfig:: { - new():: {}, - mixin:: { - // clientIP contains the configurations of Client IP based session affinity. - clientIp:: { - local __clientIpMixin(clientIp) = { clientIp+: clientIp }, - mixinInstance(clientIp):: __clientIpMixin(clientIp), - // timeoutSeconds specifies the seconds of ClientIP type session sticky time. The value must be >0 && <=86400(for 1 day) if ServiceAffinity == "ClientIP". Default value is 10800(for 3 hours). - withTimeoutSeconds(timeoutSeconds):: self + __clientIpMixin({ timeoutSeconds: timeoutSeconds }), - }, - clientIPType:: hidden.core.v1.clientIpConfig, - }, - }, - // Represents a StorageOS persistent volume resource. - storageOsPersistentVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace. - withVolumeName(volumeName):: self + { volumeName: volumeName }, - // VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to "default" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created. - withVolumeNamespace(volumeNamespace):: self + { volumeNamespace: volumeNamespace }, - mixin:: { - // SecretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: "spec.containers{name}" (where "name" refers to the name of the container that triggered the event) or if no container name is specified "spec.containers[2]" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. - withFieldPath(fieldPath):: self + __secretRefMixin({ fieldPath: fieldPath }), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - // Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/ - withNamespace(namespace):: self + __secretRefMixin({ namespace: namespace }), - // Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#concurrency-control-and-consistency - withResourceVersion(resourceVersion):: self + __secretRefMixin({ resourceVersion: resourceVersion }), - // UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids - withUid(uid):: self + __secretRefMixin({ uid: uid }), - }, - secretRefType:: hidden.core.v1.objectReference, - }, - }, - // Represents a StorageOS persistent volume resource. - storageOsVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace. - withVolumeName(volumeName):: self + { volumeName: volumeName }, - // VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to "default" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created. - withVolumeNamespace(volumeNamespace):: self + { volumeNamespace: volumeNamespace }, - mixin:: { - // SecretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted. - secretRef:: { - local __secretRefMixin(secretRef) = { secretRef+: secretRef }, - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - }, - // Sysctl defines a kernel parameter to be set - sysctl:: { - new():: {}, - // Name of a property to set - withName(name):: self + { name: name }, - // Value of a property to set - withValue(value):: self + { value: value }, - mixin:: {}, - }, - // The node this Taint is attached to has the "effect" on any pod that does not tolerate the Taint. - taint:: { - new():: {}, - // Required. The effect of the taint on pods that do not tolerate the taint. Valid effects are NoSchedule, PreferNoSchedule and NoExecute. - withEffect(effect):: self + { effect: effect }, - // Required. The taint key to be applied to a node. - withKey(key):: self + { key: key }, - // TimeAdded represents the time at which the taint was added. It is only written for NoExecute taints. - withTimeAdded(timeAdded):: self + { timeAdded: timeAdded }, - // Required. The taint value corresponding to the taint key. - withValue(value):: self + { value: value }, - mixin:: {}, - }, - // TCPSocketAction describes an action based on opening a socket - tcpSocketAction:: { - new():: {}, - // Optional: Host name to connect to, defaults to the pod IP. - withHost(host):: self + { host: host }, - // Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME. - withPort(port):: self + { port: port }, - mixin:: {}, - }, - // The pod this Toleration is attached to tolerates any taint that matches the triple using the matching operator . - toleration:: { - new():: {}, - // Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute. - withEffect(effect):: self + { effect: effect }, - // Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys. - withKey(key):: self + { key: key }, - // Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category. - withOperator(operator):: self + { operator: operator }, - // TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system. - withTolerationSeconds(tolerationSeconds):: self + { tolerationSeconds: tolerationSeconds }, - // Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string. - withValue(value):: self + { value: value }, - mixin:: {}, - }, - // A topology selector requirement is a selector that matches given label. This is an alpha feature and may change in the future. - topologySelectorLabelRequirement:: { - new():: {}, - // The label key that the selector applies to. - withKey(key):: self + { key: key }, - // An array of string values. One value must match the label to be selected. Each entry in Values is ORed. - withValues(values):: self + if std.type(values) == 'array' then { values: values } else { values: [values] }, - // An array of string values. One value must match the label to be selected. Each entry in Values is ORed. - withValuesMixin(values):: self + if std.type(values) == 'array' then { values+: values } else { values+: [values] }, - mixin:: {}, - }, - // A topology selector term represents the result of label queries. A null or empty topology selector term matches no objects. The requirements of them are ANDed. It provides a subset of functionality as NodeSelectorTerm. This is an alpha feature and may change in the future. - topologySelectorTerm:: { - new():: {}, - // A list of topology selector requirements by labels. - withMatchLabelExpressions(matchLabelExpressions):: self + if std.type(matchLabelExpressions) == 'array' then { matchLabelExpressions: matchLabelExpressions } else { matchLabelExpressions: [matchLabelExpressions] }, - // A list of topology selector requirements by labels. - withMatchLabelExpressionsMixin(matchLabelExpressions):: self + if std.type(matchLabelExpressions) == 'array' then { matchLabelExpressions+: matchLabelExpressions } else { matchLabelExpressions+: [matchLabelExpressions] }, - matchLabelExpressionsType:: hidden.core.v1.topologySelectorLabelRequirement, - mixin:: {}, - }, - // TypedLocalObjectReference contains enough information to let you locate the typed referenced object inside the same namespace. - typedLocalObjectReference:: { - new():: {}, - // APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required. - withApiGroup(apiGroup):: self + { apiGroup: apiGroup }, - // Kind is the type of resource being referenced - withKind(kind):: self + { kind: kind }, - // Name is the name of resource being referenced - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Volume represents a named volume in a pod that may be accessed by any container in the pod. - volume:: { - fromConfigMap(name='', configMapName='', configMapItems=''):: self.withName(name) + self.mixin.configMap.withItems(configMapItems).withName(configMapName), - fromEmptyDir(name='', emptyDir={}):: self.withName(name) + self.mixin.emptyDir.mixinInstance(emptyDir), - fromPersistentVolumeClaim(name='', emptyDir=''):: self.withName(name) + self.mixin.persistentVolumeClaim.withClaimName(emptyDir), - fromHostPath(name='', hostPath=''):: self.withName(name) + self.mixin.hostPath.withPath(hostPath), - fromSecret(name='', secretName=''):: self.withName(name) + self.mixin.secret.withSecretName(secretName), - // Volume's name. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + { name: name }, - mixin:: { - // AWSElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - awsElasticBlockStore:: { - local __awsElasticBlockStoreMixin(awsElasticBlockStore) = { awsElasticBlockStore+: awsElasticBlockStore }, - mixinInstance(awsElasticBlockStore):: __awsElasticBlockStoreMixin(awsElasticBlockStore), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withFsType(fsType):: self + __awsElasticBlockStoreMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). - withPartition(partition):: self + __awsElasticBlockStoreMixin({ partition: partition }), - // Specify "true" to force and set the ReadOnly property in VolumeMounts to "true". If omitted, the default is "false". More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withReadOnly(readOnly):: self + __awsElasticBlockStoreMixin({ readOnly: readOnly }), - // Unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore - withVolumeId(volumeId):: self + __awsElasticBlockStoreMixin({ volumeID: volumeId }), - }, - awsElasticBlockStoreType:: hidden.core.v1.awsElasticBlockStoreVolumeSource, - // AzureDisk represents an Azure Data Disk mount on the host and bind mount to the pod. - azureDisk:: { - local __azureDiskMixin(azureDisk) = { azureDisk+: azureDisk }, - mixinInstance(azureDisk):: __azureDiskMixin(azureDisk), - // Host Caching mode: None, Read Only, Read Write. - withCachingMode(cachingMode):: self + __azureDiskMixin({ cachingMode: cachingMode }), - // The Name of the data disk in the blob storage - withDiskName(diskName):: self + __azureDiskMixin({ diskName: diskName }), - // The URI the data disk in the blob storage - withDiskUri(diskUri):: self + __azureDiskMixin({ diskURI: diskUri }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __azureDiskMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureDiskMixin({ readOnly: readOnly }), - }, - azureDiskType:: hidden.core.v1.azureDiskVolumeSource, - // AzureFile represents an Azure File Service mount on the host and bind mount to the pod. - azureFile:: { - local __azureFileMixin(azureFile) = { azureFile+: azureFile }, - mixinInstance(azureFile):: __azureFileMixin(azureFile), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __azureFileMixin({ readOnly: readOnly }), - // the name of secret that contains Azure Storage Account Name and Key - withSecretName(secretName):: self + __azureFileMixin({ secretName: secretName }), - // Share Name - withShareName(shareName):: self + __azureFileMixin({ shareName: shareName }), - }, - azureFileType:: hidden.core.v1.azureFileVolumeSource, - // CephFS represents a Ceph FS mount on the host that shares a pod's lifetime - cephfs:: { - local __cephfsMixin(cephfs) = { cephfs+: cephfs }, - mixinInstance(cephfs):: __cephfsMixin(cephfs), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors: monitors }) else __cephfsMixin({ monitors: [monitors] }), - // Required: Monitors is a collection of Ceph monitors More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __cephfsMixin({ monitors+: monitors }) else __cephfsMixin({ monitors+: [monitors] }), - // Optional: Used as the mounted root, rather than the full Ceph tree, default is / - withPath(path):: self + __cephfsMixin({ path: path }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withReadOnly(readOnly):: self + __cephfsMixin({ readOnly: readOnly }), - // Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withSecretFile(secretFile):: self + __cephfsMixin({ secretFile: secretFile }), - // Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __cephfsMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // Optional: User is the rados user name, default is admin More info: https://releases.k8s.io/HEAD/examples/volumes/cephfs/README.md#how-to-use-it - withUser(user):: self + __cephfsMixin({ user: user }), - }, - cephfsType:: hidden.core.v1.cephFsVolumeSource, - // Cinder represents a cinder volume attached and mounted on kubelets host machine More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - cinder:: { - local __cinderMixin(cinder) = { cinder+: cinder }, - mixinInstance(cinder):: __cinderMixin(cinder), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withFsType(fsType):: self + __cinderMixin({ fsType: fsType }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withReadOnly(readOnly):: self + __cinderMixin({ readOnly: readOnly }), - // Optional: points to a secret object containing parameters used to connect to OpenStack. - secretRef:: { - local __secretRefMixin(secretRef) = __cinderMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // volume id used to identify the volume in cinder More info: https://releases.k8s.io/HEAD/examples/mysql-cinder-pd/README.md - withVolumeId(volumeId):: self + __cinderMixin({ volumeID: volumeId }), - }, - cinderType:: hidden.core.v1.cinderVolumeSource, - // ConfigMap represents a configMap that should populate this volume - configMap:: { - local __configMapMixin(configMap) = { configMap+: configMap }, - mixinInstance(configMap):: __configMapMixin(configMap), - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + __configMapMixin({ defaultMode: defaultMode }), - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then __configMapMixin({ items: items }) else __configMapMixin({ items: [items] }), - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then __configMapMixin({ items+: items }) else __configMapMixin({ items+: [items] }), - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __configMapMixin({ name: name }), - // Specify whether the ConfigMap or it's keys must be defined - withOptional(optional):: self + __configMapMixin({ optional: optional }), - }, - configMapType:: hidden.core.v1.configMapVolumeSource, - // CSI (Container Storage Interface) represents storage that is handled by an external CSI driver (Alpha feature). - csi:: { - local __csiMixin(csi) = { csi+: csi }, - mixinInstance(csi):: __csiMixin(csi), - // Driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster. - withDriver(driver):: self + __csiMixin({ driver: driver }), - // Filesystem type to mount. Ex. "ext4", "xfs", "ntfs". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply. - withFsType(fsType):: self + __csiMixin({ fsType: fsType }), - // NodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed. - nodePublishSecretRef:: { - local __nodePublishSecretRefMixin(nodePublishSecretRef) = __csiMixin({ nodePublishSecretRef+: nodePublishSecretRef }), - mixinInstance(nodePublishSecretRef):: __nodePublishSecretRefMixin(nodePublishSecretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __nodePublishSecretRefMixin({ name: name }), - }, - nodePublishSecretRefType:: hidden.core.v1.localObjectReference, - // Specifies a read-only configuration for the volume. Defaults to false (read/write). - withReadOnly(readOnly):: self + __csiMixin({ readOnly: readOnly }), - // VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values. - withVolumeAttributes(volumeAttributes):: self + __csiMixin({ volumeAttributes: volumeAttributes }), - // VolumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values. - withVolumeAttributesMixin(volumeAttributes):: self + __csiMixin({ volumeAttributes+: volumeAttributes }), - }, - csiType:: hidden.core.v1.csiVolumeSource, - // DownwardAPI represents downward API about the pod that should populate this volume - downwardApi:: { - local __downwardApiMixin(downwardApi) = { downwardApi+: downwardApi }, - mixinInstance(downwardApi):: __downwardApiMixin(downwardApi), - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + __downwardApiMixin({ defaultMode: defaultMode }), - // Items is a list of downward API volume file - withItems(items):: self + if std.type(items) == 'array' then __downwardApiMixin({ items: items }) else __downwardApiMixin({ items: [items] }), - // Items is a list of downward API volume file - withItemsMixin(items):: self + if std.type(items) == 'array' then __downwardApiMixin({ items+: items }) else __downwardApiMixin({ items+: [items] }), - itemsType:: hidden.core.v1.downwardApiVolumeFile, - }, - downwardAPIType:: hidden.core.v1.downwardApiVolumeSource, - // EmptyDir represents a temporary directory that shares a pod's lifetime. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir - emptyDir:: { - local __emptyDirMixin(emptyDir) = { emptyDir+: emptyDir }, - mixinInstance(emptyDir):: __emptyDirMixin(emptyDir), - // What type of storage medium should back this directory. The default is "" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir - withMedium(medium):: self + __emptyDirMixin({ medium: medium }), - // Total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: http://kubernetes.io/docs/user-guide/volumes#emptydir - sizeLimit:: { - local __sizeLimitMixin(sizeLimit) = __emptyDirMixin({ sizeLimit+: sizeLimit }), - mixinInstance(sizeLimit):: __sizeLimitMixin(sizeLimit), - }, - sizeLimitType:: hidden.core.resource.quantity, - }, - emptyDirType:: hidden.core.v1.emptyDirVolumeSource, - // FC represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod. - fc:: { - local __fcMixin(fc) = { fc+: fc }, - mixinInstance(fc):: __fcMixin(fc), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __fcMixin({ fsType: fsType }), - // Optional: FC target lun number - withLun(lun):: self + __fcMixin({ lun: lun }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __fcMixin({ readOnly: readOnly }), - // Optional: FC target worldwide names (WWNs) - withTargetWwns(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs: targetWwns }) else __fcMixin({ targetWWNs: [targetWwns] }), - // Optional: FC target worldwide names (WWNs) - withTargetWwnsMixin(targetWwns):: self + if std.type(targetWwns) == 'array' then __fcMixin({ targetWWNs+: targetWwns }) else __fcMixin({ targetWWNs+: [targetWwns] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwids(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids: wwids }) else __fcMixin({ wwids: [wwids] }), - // Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously. - withWwidsMixin(wwids):: self + if std.type(wwids) == 'array' then __fcMixin({ wwids+: wwids }) else __fcMixin({ wwids+: [wwids] }), - }, - fcType:: hidden.core.v1.fcVolumeSource, - // FlexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin. - flexVolume:: { - local __flexVolumeMixin(flexVolume) = { flexVolume+: flexVolume }, - mixinInstance(flexVolume):: __flexVolumeMixin(flexVolume), - // Driver is the name of the driver to use for this volume. - withDriver(driver):: self + __flexVolumeMixin({ driver: driver }), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". The default filesystem depends on FlexVolume script. - withFsType(fsType):: self + __flexVolumeMixin({ fsType: fsType }), - // Optional: Extra command options if any. - withOptions(options):: self + __flexVolumeMixin({ options: options }), - // Optional: Extra command options if any. - withOptionsMixin(options):: self + __flexVolumeMixin({ options+: options }), - // Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __flexVolumeMixin({ readOnly: readOnly }), - // Optional: SecretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts. - secretRef:: { - local __secretRefMixin(secretRef) = __flexVolumeMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - }, - flexVolumeType:: hidden.core.v1.flexVolumeSource, - // Flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running - flocker:: { - local __flockerMixin(flocker) = { flocker+: flocker }, - mixinInstance(flocker):: __flockerMixin(flocker), - // Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated - withDatasetName(datasetName):: self + __flockerMixin({ datasetName: datasetName }), - // UUID of the dataset. This is unique identifier of a Flocker dataset - withDatasetUuid(datasetUuid):: self + __flockerMixin({ datasetUUID: datasetUuid }), - }, - flockerType:: hidden.core.v1.flockerVolumeSource, - // GCEPersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - gcePersistentDisk:: { - local __gcePersistentDiskMixin(gcePersistentDisk) = { gcePersistentDisk+: gcePersistentDisk }, - mixinInstance(gcePersistentDisk):: __gcePersistentDiskMixin(gcePersistentDisk), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withFsType(fsType):: self + __gcePersistentDiskMixin({ fsType: fsType }), - // The partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as "1". Similarly, the volume partition for /dev/sda is "0" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPartition(partition):: self + __gcePersistentDiskMixin({ partition: partition }), - // Unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withPdName(pdName):: self + __gcePersistentDiskMixin({ pdName: pdName }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk - withReadOnly(readOnly):: self + __gcePersistentDiskMixin({ readOnly: readOnly }), - }, - gcePersistentDiskType:: hidden.core.v1.gcePersistentDiskVolumeSource, - // GitRepo represents a git repository at a particular revision. DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container. - gitRepo:: { - local __gitRepoMixin(gitRepo) = { gitRepo+: gitRepo }, - mixinInstance(gitRepo):: __gitRepoMixin(gitRepo), - // Target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name. - withDirectory(directory):: self + __gitRepoMixin({ directory: directory }), - // Repository URL - withRepository(repository):: self + __gitRepoMixin({ repository: repository }), - // Commit hash for the specified revision. - withRevision(revision):: self + __gitRepoMixin({ revision: revision }), - }, - gitRepoType:: hidden.core.v1.gitRepoVolumeSource, - // Glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md - glusterfs:: { - local __glusterfsMixin(glusterfs) = { glusterfs+: glusterfs }, - mixinInstance(glusterfs):: __glusterfsMixin(glusterfs), - // EndpointsName is the endpoint name that details Glusterfs topology. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withEndpoints(endpoints):: self + __glusterfsMixin({ endpoints: endpoints }), - // Path is the Glusterfs volume path. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withPath(path):: self + __glusterfsMixin({ path: path }), - // ReadOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/glusterfs/README.md#create-a-pod - withReadOnly(readOnly):: self + __glusterfsMixin({ readOnly: readOnly }), - }, - glusterfsType:: hidden.core.v1.glusterfsVolumeSource, - // HostPath represents a pre-existing file or directory on the host machine that is directly exposed to the container. This is generally used for system agents or other privileged things that are allowed to see the host machine. Most containers will NOT need this. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - hostPath:: { - local __hostPathMixin(hostPath) = { hostPath+: hostPath }, - mixinInstance(hostPath):: __hostPathMixin(hostPath), - // Path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withPath(path):: self + __hostPathMixin({ path: path }), - // Type for HostPath Volume Defaults to "" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath - withType(type):: self + __hostPathMixin({ type: type }), - }, - hostPathType:: hidden.core.v1.hostPathVolumeSource, - // ISCSI represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://releases.k8s.io/HEAD/examples/volumes/iscsi/README.md - iscsi:: { - local __iscsiMixin(iscsi) = { iscsi+: iscsi }, - mixinInstance(iscsi):: __iscsiMixin(iscsi), - // whether support iSCSI Discovery CHAP authentication - withChapAuthDiscovery(chapAuthDiscovery):: self + __iscsiMixin({ chapAuthDiscovery: chapAuthDiscovery }), - // whether support iSCSI Session CHAP authentication - withChapAuthSession(chapAuthSession):: self + __iscsiMixin({ chapAuthSession: chapAuthSession }), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi - withFsType(fsType):: self + __iscsiMixin({ fsType: fsType }), - // Custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection. - withInitiatorName(initiatorName):: self + __iscsiMixin({ initiatorName: initiatorName }), - // Target iSCSI Qualified Name. - withIqn(iqn):: self + __iscsiMixin({ iqn: iqn }), - // iSCSI Interface Name that uses an iSCSI transport. Defaults to 'default' (tcp). - withIscsiInterface(iscsiInterface):: self + __iscsiMixin({ iscsiInterface: iscsiInterface }), - // iSCSI Target Lun number. - withLun(lun):: self + __iscsiMixin({ lun: lun }), - // iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortals(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals: portals }) else __iscsiMixin({ portals: [portals] }), - // iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withPortalsMixin(portals):: self + if std.type(portals) == 'array' then __iscsiMixin({ portals+: portals }) else __iscsiMixin({ portals+: [portals] }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. - withReadOnly(readOnly):: self + __iscsiMixin({ readOnly: readOnly }), - // CHAP Secret for iSCSI target and initiator authentication - secretRef:: { - local __secretRefMixin(secretRef) = __iscsiMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260). - withTargetPortal(targetPortal):: self + __iscsiMixin({ targetPortal: targetPortal }), - }, - iscsiType:: hidden.core.v1.iscsiVolumeSource, - // NFS represents an NFS mount on the host that shares a pod's lifetime More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - nfs:: { - local __nfsMixin(nfs) = { nfs+: nfs }, - mixinInstance(nfs):: __nfsMixin(nfs), - // Path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withPath(path):: self + __nfsMixin({ path: path }), - // ReadOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withReadOnly(readOnly):: self + __nfsMixin({ readOnly: readOnly }), - // Server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs - withServer(server):: self + __nfsMixin({ server: server }), - }, - nfsType:: hidden.core.v1.nfsVolumeSource, - // PersistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - persistentVolumeClaim:: { - local __persistentVolumeClaimMixin(persistentVolumeClaim) = { persistentVolumeClaim+: persistentVolumeClaim }, - mixinInstance(persistentVolumeClaim):: __persistentVolumeClaimMixin(persistentVolumeClaim), - // ClaimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims - withClaimName(claimName):: self + __persistentVolumeClaimMixin({ claimName: claimName }), - // Will force the ReadOnly setting in VolumeMounts. Default false. - withReadOnly(readOnly):: self + __persistentVolumeClaimMixin({ readOnly: readOnly }), - }, - persistentVolumeClaimType:: hidden.core.v1.persistentVolumeClaimVolumeSource, - // PhotonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine - photonPersistentDisk:: { - local __photonPersistentDiskMixin(photonPersistentDisk) = { photonPersistentDisk+: photonPersistentDisk }, - mixinInstance(photonPersistentDisk):: __photonPersistentDiskMixin(photonPersistentDisk), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __photonPersistentDiskMixin({ fsType: fsType }), - // ID that identifies Photon Controller persistent disk - withPdId(pdId):: self + __photonPersistentDiskMixin({ pdID: pdId }), - }, - photonPersistentDiskType:: hidden.core.v1.photonPersistentDiskVolumeSource, - // PortworxVolume represents a portworx volume attached and mounted on kubelets host machine - portworxVolume:: { - local __portworxVolumeMixin(portworxVolume) = { portworxVolume+: portworxVolume }, - mixinInstance(portworxVolume):: __portworxVolumeMixin(portworxVolume), - // FSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __portworxVolumeMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __portworxVolumeMixin({ readOnly: readOnly }), - // VolumeID uniquely identifies a Portworx volume - withVolumeId(volumeId):: self + __portworxVolumeMixin({ volumeID: volumeId }), - }, - portworxVolumeType:: hidden.core.v1.portworxVolumeSource, - // Items for all in one resources secrets, configmaps, and downward API - projected:: { - local __projectedMixin(projected) = { projected+: projected }, - mixinInstance(projected):: __projectedMixin(projected), - // Mode bits to use on created files by default. Must be a value between 0 and 0777. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + __projectedMixin({ defaultMode: defaultMode }), - // list of volume projections - withSources(sources):: self + if std.type(sources) == 'array' then __projectedMixin({ sources: sources }) else __projectedMixin({ sources: [sources] }), - // list of volume projections - withSourcesMixin(sources):: self + if std.type(sources) == 'array' then __projectedMixin({ sources+: sources }) else __projectedMixin({ sources+: [sources] }), - sourcesType:: hidden.core.v1.volumeProjection, - }, - projectedType:: hidden.core.v1.projectedVolumeSource, - // Quobyte represents a Quobyte mount on the host that shares a pod's lifetime - quobyte:: { - local __quobyteMixin(quobyte) = { quobyte+: quobyte }, - mixinInstance(quobyte):: __quobyteMixin(quobyte), - // Group to map volume access to Default is no group - withGroup(group):: self + __quobyteMixin({ group: group }), - // ReadOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false. - withReadOnly(readOnly):: self + __quobyteMixin({ readOnly: readOnly }), - // Registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes - withRegistry(registry):: self + __quobyteMixin({ registry: registry }), - // Tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin - withTenant(tenant):: self + __quobyteMixin({ tenant: tenant }), - // User to map volume access to Defaults to serivceaccount user - withUser(user):: self + __quobyteMixin({ user: user }), - // Volume is a string that references an already created Quobyte volume by name. - withVolume(volume):: self + __quobyteMixin({ volume: volume }), - }, - quobyteType:: hidden.core.v1.quobyteVolumeSource, - // RBD represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md - rbd:: { - local __rbdMixin(rbd) = { rbd+: rbd }, - mixinInstance(rbd):: __rbdMixin(rbd), - // Filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd - withFsType(fsType):: self + __rbdMixin({ fsType: fsType }), - // The rados image name. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withImage(image):: self + __rbdMixin({ image: image }), - // Keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withKeyring(keyring):: self + __rbdMixin({ keyring: keyring }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitors(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors: monitors }) else __rbdMixin({ monitors: [monitors] }), - // A collection of Ceph monitors. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withMonitorsMixin(monitors):: self + if std.type(monitors) == 'array' then __rbdMixin({ monitors+: monitors }) else __rbdMixin({ monitors+: [monitors] }), - // The rados pool name. Default is rbd. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withPool(pool):: self + __rbdMixin({ pool: pool }), - // ReadOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withReadOnly(readOnly):: self + __rbdMixin({ readOnly: readOnly }), - // SecretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - secretRef:: { - local __secretRefMixin(secretRef) = __rbdMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // The rados user name. Default is admin. More info: https://releases.k8s.io/HEAD/examples/volumes/rbd/README.md#how-to-use-it - withUser(user):: self + __rbdMixin({ user: user }), - }, - rbdType:: hidden.core.v1.rbdVolumeSource, - // ScaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes. - scaleIo:: { - local __scaleIoMixin(scaleIo) = { scaleIo+: scaleIo }, - mixinInstance(scaleIo):: __scaleIoMixin(scaleIo), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Default is "xfs". - withFsType(fsType):: self + __scaleIoMixin({ fsType: fsType }), - // The host address of the ScaleIO API Gateway. - withGateway(gateway):: self + __scaleIoMixin({ gateway: gateway }), - // The name of the ScaleIO Protection Domain for the configured storage. - withProtectionDomain(protectionDomain):: self + __scaleIoMixin({ protectionDomain: protectionDomain }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __scaleIoMixin({ readOnly: readOnly }), - // SecretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail. - secretRef:: { - local __secretRefMixin(secretRef) = __scaleIoMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // Flag to enable/disable SSL communication with Gateway, default false - withSslEnabled(sslEnabled):: self + __scaleIoMixin({ sslEnabled: sslEnabled }), - // Indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned. - withStorageMode(storageMode):: self + __scaleIoMixin({ storageMode: storageMode }), - // The ScaleIO Storage Pool associated with the protection domain. - withStoragePool(storagePool):: self + __scaleIoMixin({ storagePool: storagePool }), - // The name of the storage system as configured in ScaleIO. - withSystem(system):: self + __scaleIoMixin({ system: system }), - // The name of a volume already created in the ScaleIO system that is associated with this volume source. - withVolumeName(volumeName):: self + __scaleIoMixin({ volumeName: volumeName }), - }, - scaleIOType:: hidden.core.v1.scaleIoVolumeSource, - // Secret represents a secret that should populate this volume. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret - secret:: { - local __secretMixin(secret) = { secret+: secret }, - mixinInstance(secret):: __secretMixin(secret), - // Optional: mode bits to use on created files by default. Must be a value between 0 and 0777. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set. - withDefaultMode(defaultMode):: self + __secretMixin({ defaultMode: defaultMode }), - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then __secretMixin({ items: items }) else __secretMixin({ items: [items] }), - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then __secretMixin({ items+: items }) else __secretMixin({ items+: [items] }), - itemsType:: hidden.core.v1.keyToPath, - // Specify whether the Secret or it's keys must be defined - withOptional(optional):: self + __secretMixin({ optional: optional }), - // Name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret - withSecretName(secretName):: self + __secretMixin({ secretName: secretName }), - }, - secretType:: hidden.core.v1.secretVolumeSource, - // StorageOS represents a StorageOS volume attached and mounted on Kubernetes nodes. - storageos:: { - local __storageosMixin(storageos) = { storageos+: storageos }, - mixinInstance(storageos):: __storageosMixin(storageos), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __storageosMixin({ fsType: fsType }), - // Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. - withReadOnly(readOnly):: self + __storageosMixin({ readOnly: readOnly }), - // SecretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted. - secretRef:: { - local __secretRefMixin(secretRef) = __storageosMixin({ secretRef+: secretRef }), - mixinInstance(secretRef):: __secretRefMixin(secretRef), - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretRefMixin({ name: name }), - }, - secretRefType:: hidden.core.v1.localObjectReference, - // VolumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace. - withVolumeName(volumeName):: self + __storageosMixin({ volumeName: volumeName }), - // VolumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to "default" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created. - withVolumeNamespace(volumeNamespace):: self + __storageosMixin({ volumeNamespace: volumeNamespace }), - }, - storageosType:: hidden.core.v1.storageOsVolumeSource, - // VsphereVolume represents a vSphere volume attached and mounted on kubelets host machine - vsphereVolume:: { - local __vsphereVolumeMixin(vsphereVolume) = { vsphereVolume+: vsphereVolume }, - mixinInstance(vsphereVolume):: __vsphereVolumeMixin(vsphereVolume), - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + __vsphereVolumeMixin({ fsType: fsType }), - // Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName. - withStoragePolicyId(storagePolicyId):: self + __vsphereVolumeMixin({ storagePolicyID: storagePolicyId }), - // Storage Policy Based Management (SPBM) profile name. - withStoragePolicyName(storagePolicyName):: self + __vsphereVolumeMixin({ storagePolicyName: storagePolicyName }), - // Path that identifies vSphere volume vmdk - withVolumePath(volumePath):: self + __vsphereVolumeMixin({ volumePath: volumePath }), - }, - vsphereVolumeType:: hidden.core.v1.vsphereVirtualDiskVolumeSource, - }, - }, - // volumeDevice describes a mapping of a raw block device within a container. - volumeDevice:: { - new():: {}, - // devicePath is the path inside of the container that the device will be mapped to. - withDevicePath(devicePath):: self + { devicePath: devicePath }, - // name must match the name of a persistentVolumeClaim in the pod - withName(name):: self + { name: name }, - mixin:: {}, - }, - // VolumeMount describes a mounting of a Volume within a container. - volumeMount:: { - new(name='', mountPath='', readOnly=false):: self.withMountPath(mountPath).withName(name).withReadOnly(readOnly), - // Path within the container at which the volume should be mounted. Must not contain ':'. - withMountPath(mountPath):: self + { mountPath: mountPath }, - // mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10. - withMountPropagation(mountPropagation):: self + { mountPropagation: mountPropagation }, - // This must match the Name of a Volume. - withName(name):: self + { name: name }, - // Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - // Path within the volume from which the container's volume should be mounted. Defaults to "" (volume's root). - withSubPath(subPath):: self + { subPath: subPath }, - // Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to "" (volume's root). SubPathExpr and SubPath are mutually exclusive. This field is alpha in 1.14. - withSubPathExpr(subPathExpr):: self + { subPathExpr: subPathExpr }, - mixin:: {}, - }, - // VolumeNodeAffinity defines constraints that limit what nodes this volume can be accessed from. - volumeNodeAffinity:: { - new():: {}, - mixin:: { - // Required specifies hard node constraints that must be met. - required:: { - local __requiredMixin(required) = { required+: required }, - mixinInstance(required):: __requiredMixin(required), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredType:: hidden.core.v1.nodeSelector, - }, - }, - // Projection that may be projected along with other supported volume types - volumeProjection:: { - new():: {}, - mixin:: { - // information about the configMap data to project - configMap:: { - local __configMapMixin(configMap) = { configMap+: configMap }, - mixinInstance(configMap):: __configMapMixin(configMap), - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then __configMapMixin({ items: items }) else __configMapMixin({ items: [items] }), - // If unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then __configMapMixin({ items+: items }) else __configMapMixin({ items+: [items] }), - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __configMapMixin({ name: name }), - // Specify whether the ConfigMap or it's keys must be defined - withOptional(optional):: self + __configMapMixin({ optional: optional }), - }, - configMapType:: hidden.core.v1.configMapProjection, - // information about the downwardAPI data to project - downwardApi:: { - local __downwardApiMixin(downwardApi) = { downwardApi+: downwardApi }, - mixinInstance(downwardApi):: __downwardApiMixin(downwardApi), - // Items is a list of DownwardAPIVolume file - withItems(items):: self + if std.type(items) == 'array' then __downwardApiMixin({ items: items }) else __downwardApiMixin({ items: [items] }), - // Items is a list of DownwardAPIVolume file - withItemsMixin(items):: self + if std.type(items) == 'array' then __downwardApiMixin({ items+: items }) else __downwardApiMixin({ items+: [items] }), - itemsType:: hidden.core.v1.downwardApiVolumeFile, - }, - downwardAPIType:: hidden.core.v1.downwardApiProjection, - // information about the secret data to project - secret:: { - local __secretMixin(secret) = { secret+: secret }, - mixinInstance(secret):: __secretMixin(secret), - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItems(items):: self + if std.type(items) == 'array' then __secretMixin({ items: items }) else __secretMixin({ items: [items] }), - // If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'. - withItemsMixin(items):: self + if std.type(items) == 'array' then __secretMixin({ items+: items }) else __secretMixin({ items+: [items] }), - itemsType:: hidden.core.v1.keyToPath, - // Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names - withName(name):: self + __secretMixin({ name: name }), - // Specify whether the Secret or its key must be defined - withOptional(optional):: self + __secretMixin({ optional: optional }), - }, - secretType:: hidden.core.v1.secretProjection, - // information about the serviceAccountToken data to project - serviceAccountToken:: { - local __serviceAccountTokenMixin(serviceAccountToken) = { serviceAccountToken+: serviceAccountToken }, - mixinInstance(serviceAccountToken):: __serviceAccountTokenMixin(serviceAccountToken), - // Audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver. - withAudience(audience):: self + __serviceAccountTokenMixin({ audience: audience }), - // ExpirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes. - withExpirationSeconds(expirationSeconds):: self + __serviceAccountTokenMixin({ expirationSeconds: expirationSeconds }), - // Path is the path relative to the mount point of the file to project the token into. - withPath(path):: self + __serviceAccountTokenMixin({ path: path }), - }, - serviceAccountTokenType:: hidden.core.v1.serviceAccountTokenProjection, - }, - }, - // Represents a vSphere volume resource. - vsphereVirtualDiskVolumeSource:: { - new():: {}, - // Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. "ext4", "xfs", "ntfs". Implicitly inferred to be "ext4" if unspecified. - withFsType(fsType):: self + { fsType: fsType }, - // Storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName. - withStoragePolicyId(storagePolicyId):: self + { storagePolicyID: storagePolicyId }, - // Storage Policy Based Management (SPBM) profile name. - withStoragePolicyName(storagePolicyName):: self + { storagePolicyName: storagePolicyName }, - // Path that identifies vSphere volume vmdk - withVolumePath(volumePath):: self + { volumePath: volumePath }, - mixin:: {}, - }, - // The weights of all of the matched WeightedPodAffinityTerm fields are added per-node to find the most preferred node(s) - weightedPodAffinityTerm:: { - new():: {}, - // weight associated with matching the corresponding podAffinityTerm, in the range 1-100. - withWeight(weight):: self + { weight: weight }, - mixin:: { - // Required. A pod affinity term, associated with the corresponding weight. - podAffinityTerm:: { - local __podAffinityTermMixin(podAffinityTerm) = { podAffinityTerm+: podAffinityTerm }, - mixinInstance(podAffinityTerm):: __podAffinityTermMixin(podAffinityTerm), - // A label query over a set of resources, in this case pods. - labelSelector:: { - local __labelSelectorMixin(labelSelector) = __podAffinityTermMixin({ labelSelector+: labelSelector }), - mixinInstance(labelSelector):: __labelSelectorMixin(labelSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __labelSelectorMixin({ matchExpressions: matchExpressions }) else __labelSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __labelSelectorMixin({ matchExpressions+: matchExpressions }) else __labelSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __labelSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __labelSelectorMixin({ matchLabels+: matchLabels }), - }, - labelSelectorType:: hidden.meta.v1.labelSelector, - // namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means "this pod's namespace" - withNamespaces(namespaces):: self + if std.type(namespaces) == 'array' then __podAffinityTermMixin({ namespaces: namespaces }) else __podAffinityTermMixin({ namespaces: [namespaces] }), - // namespaces specifies which namespaces the labelSelector applies to (matches against); null or empty list means "this pod's namespace" - withNamespacesMixin(namespaces):: self + if std.type(namespaces) == 'array' then __podAffinityTermMixin({ namespaces+: namespaces }) else __podAffinityTermMixin({ namespaces+: [namespaces] }), - // This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed. - withTopologyKey(topologyKey):: self + __podAffinityTermMixin({ topologyKey: topologyKey }), - }, - podAffinityTermType:: hidden.core.v1.podAffinityTerm, - }, - }, - }, - }, - events:: { - v1beta1:: { - local apiVersion = { apiVersion: 'events/v1beta1' }, - // EventList is a list of Event objects. - eventList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.events.v1beta1.event, - mixin:: {}, - }, - // EventSeries contain information on series of events, i.e. thing that was/is happening continuously for some time. - eventSeries:: { - new():: {}, - // Number of occurrences in this series up to the last heartbeat time - withCount(count):: self + { count: count }, - // Time when last Event from the series was seen before last heartbeat. - withLastObservedTime(lastObservedTime):: self + { lastObservedTime: lastObservedTime }, - // Information whether this series is ongoing or finished. - withState(state):: self + { state: state }, - mixin:: {}, - }, - }, - }, - extensions:: { - v1beta1:: { - local apiVersion = { apiVersion: 'extensions/v1beta1' }, - // AllowedCSIDriver represents a single inline CSI Driver that is allowed to be used. - allowedCsiDriver:: { - new():: {}, - // Name is the registered name of the CSI driver - withName(name):: self + { name: name }, - mixin:: {}, - }, - // AllowedFlexVolume represents a single Flexvolume that is allowed to be used. Deprecated: use AllowedFlexVolume from policy API Group instead. - allowedFlexVolume:: { - new():: {}, - // driver is the name of the Flexvolume driver. - withDriver(driver):: self + { driver: driver }, - mixin:: {}, - }, - // AllowedHostPath defines the host volume conditions that will be enabled by a policy for pods to use. It requires the path prefix to be defined. Deprecated: use AllowedHostPath from policy API Group instead. - allowedHostPath:: { - new():: {}, - // pathPrefix is the path prefix that the host volume must match. It does not support `*`. Trailing slashes are trimmed when validating the path prefix with a host path. - // - // Examples: `/foo` would allow `/foo`, `/foo/` and `/foo/bar` `/foo` would not allow `/food` or `/etc/foo` - withPathPrefix(pathPrefix):: self + { pathPrefix: pathPrefix }, - // when set to true, will allow host volumes matching the pathPrefix only if all volume mounts are readOnly. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // DaemonSetCondition describes the state of a DaemonSet at a certain point. - daemonSetCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of DaemonSet condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DaemonSetList is a collection of daemon sets. - daemonSetList:: { - new():: {}, - // A list of daemon sets. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // A list of daemon sets. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.daemonSet, - mixin:: {}, - }, - // DaemonSetSpec is the specification of a daemon set. - daemonSetSpec:: { - new():: {}, - // The minimum number of seconds for which a newly created DaemonSet pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready). - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // The number of old history to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. Defaults to 10. - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - // DEPRECATED. A sequence number representing a specific generation of the template. Populated by the system. It can be set only during the creation. - withTemplateGeneration(templateGeneration):: self + { templateGeneration: templateGeneration }, - mixin:: { - // A label query over pods that are managed by the daemon set. Must match in order to be controlled. If empty, defaulted to labels on Pod template. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // An object that describes the pod that will be created. The DaemonSet will create exactly one copy of this pod on every node that matches the template's node selector (or on every node if no node selector is specified). More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - // An update strategy to replace existing DaemonSet pods with new pods. - updateStrategy:: { - local __updateStrategyMixin(updateStrategy) = { updateStrategy+: updateStrategy }, - mixinInstance(updateStrategy):: __updateStrategyMixin(updateStrategy), - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __updateStrategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDaemonSet, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is OnDelete. - withType(type):: self + __updateStrategyMixin({ type: type }), - }, - updateStrategyType:: hidden.extensions.v1beta1.daemonSetUpdateStrategy, - }, - }, - // DaemonSetStatus represents the current status of a daemon set. - daemonSetStatus:: { - new():: {}, - // Count of hash collisions for the DaemonSet. The DaemonSet controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ControllerRevision. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a DaemonSet's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a DaemonSet's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.extensions.v1beta1.daemonSetCondition, - // The number of nodes that are running at least 1 daemon pod and are supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withCurrentNumberScheduled(currentNumberScheduled):: self + { currentNumberScheduled: currentNumberScheduled }, - // The total number of nodes that should be running the daemon pod (including nodes correctly running the daemon pod). More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withDesiredNumberScheduled(desiredNumberScheduled):: self + { desiredNumberScheduled: desiredNumberScheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberAvailable(numberAvailable):: self + { numberAvailable: numberAvailable }, - // The number of nodes that are running the daemon pod, but are not supposed to run the daemon pod. More info: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ - withNumberMisscheduled(numberMisscheduled):: self + { numberMisscheduled: numberMisscheduled }, - // The number of nodes that should be running the daemon pod and have one or more of the daemon pod running and ready. - withNumberReady(numberReady):: self + { numberReady: numberReady }, - // The number of nodes that should be running the daemon pod and have none of the daemon pod running and available (ready for at least spec.minReadySeconds) - withNumberUnavailable(numberUnavailable):: self + { numberUnavailable: numberUnavailable }, - // The most recent generation observed by the daemon set controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The total number of nodes that are running updated daemon pod - withUpdatedNumberScheduled(updatedNumberScheduled):: self + { updatedNumberScheduled: updatedNumberScheduled }, - mixin:: {}, - }, - daemonSetUpdateStrategy:: { - new():: {}, - // Type of daemon set update. Can be "RollingUpdate" or "OnDelete". Default is OnDelete. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if type = "RollingUpdate". - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDaemonSet, - }, - }, - // DeploymentCondition describes the state of a deployment at a certain point. - deploymentCondition:: { - new():: {}, - // Last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // The last time this condition was updated. - withLastUpdateTime(lastUpdateTime):: self + { lastUpdateTime: lastUpdateTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of deployment condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // DeploymentList is a list of Deployments. - deploymentList:: { - new(items=''):: self.withItems(items), - // Items is the list of Deployments. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Deployments. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.deployment, - mixin:: {}, - }, - // DeploymentSpec is the specification of the desired behavior of the Deployment. - deploymentSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Indicates that the deployment is paused and will not be processed by the deployment controller. - withPaused(paused):: self + { paused: paused }, - // The maximum time in seconds for a deployment to make progress before it is considered to be failed. The deployment controller will continue to process failed deployments and a condition with a ProgressDeadlineExceeded reason will be surfaced in the deployment status. Note that progress will not be estimated during the time a deployment is paused. This is set to the max value of int32 (i.e. 2147483647) by default, which means "no deadline". - withProgressDeadlineSeconds(progressDeadlineSeconds):: self + { progressDeadlineSeconds: progressDeadlineSeconds }, - // Number of desired pods. This is a pointer to distinguish between explicit zero and not specified. Defaults to 1. - withReplicas(replicas):: self + { replicas: replicas }, - // The number of old ReplicaSets to retain to allow rollback. This is a pointer to distinguish between explicit zero and not specified. This is set to the max value of int32 (i.e. 2147483647) by default, which means "retaining all old RelicaSets". - withRevisionHistoryLimit(revisionHistoryLimit):: self + { revisionHistoryLimit: revisionHistoryLimit }, - mixin:: { - // DEPRECATED. The config this deployment is rolling back to. Will be cleared after rollback is done. - rollbackTo:: { - local __rollbackToMixin(rollbackTo) = { rollbackTo+: rollbackTo }, - mixinInstance(rollbackTo):: __rollbackToMixin(rollbackTo), - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + __rollbackToMixin({ revision: revision }), - }, - rollbackToType:: hidden.extensions.v1beta1.rollbackConfig, - // Label selector for pods. Existing ReplicaSets whose pods are selected by this will be the ones affected by this deployment. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // The deployment strategy to use to replace existing pods with new ones. - strategy:: { - local __strategyMixin(strategy) = { strategy+: strategy }, - mixinInstance(strategy):: __strategyMixin(strategy), - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = __strategyMixin({ rollingUpdate+: rollingUpdate }), - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. By default, a value of 1 is used. Example: when this is set to 30%, the new RC can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new RC can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. By default, a fixed value of 1 is used. Example: when this is set to 30%, the old RC can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old RC can be scaled down further, followed by scaling up the new RC, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDeployment, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + __strategyMixin({ type: type }), - }, - strategyType:: hidden.extensions.v1beta1.deploymentStrategy, - // Template describes the pods that will be created. - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // DeploymentStatus is the most recently observed status of the Deployment. - deploymentStatus:: { - new():: {}, - // Total number of available pods (ready for at least minReadySeconds) targeted by this deployment. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Count of hash collisions for the Deployment. The Deployment controller uses this field as a collision avoidance mechanism when it needs to create the name for the newest ReplicaSet. - withCollisionCount(collisionCount):: self + { collisionCount: collisionCount }, - // Represents the latest available observations of a deployment's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a deployment's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.extensions.v1beta1.deploymentCondition, - // The generation observed by the deployment controller. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // Total number of ready pods targeted by this deployment. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Total number of non-terminated pods targeted by this deployment (their labels match the selector). - withReplicas(replicas):: self + { replicas: replicas }, - // Total number of unavailable pods targeted by this deployment. This is the total number of pods that are still required for the deployment to have 100% available capacity. They may either be pods that are running but not yet available or pods that still have not been created. - withUnavailableReplicas(unavailableReplicas):: self + { unavailableReplicas: unavailableReplicas }, - // Total number of non-terminated pods targeted by this deployment that have the desired template spec. - withUpdatedReplicas(updatedReplicas):: self + { updatedReplicas: updatedReplicas }, - mixin:: {}, - }, - // DeploymentStrategy describes how to replace existing pods with new ones. - deploymentStrategy:: { - new():: {}, - // Type of deployment. Can be "Recreate" or "RollingUpdate". Default is RollingUpdate. - withType(type):: self + { type: type }, - mixin:: { - // Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. - rollingUpdate:: { - local __rollingUpdateMixin(rollingUpdate) = { rollingUpdate+: rollingUpdate }, - mixinInstance(rollingUpdate):: __rollingUpdateMixin(rollingUpdate), - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. By default, a value of 1 is used. Example: when this is set to 30%, the new RC can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new RC can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + __rollingUpdateMixin({ maxSurge: maxSurge }), - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. By default, a fixed value of 1 is used. Example: when this is set to 30%, the old RC can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old RC can be scaled down further, followed by scaling up the new RC, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + __rollingUpdateMixin({ maxUnavailable: maxUnavailable }), - }, - rollingUpdateType:: hidden.extensions.v1beta1.rollingUpdateDeployment, - }, - }, - // FSGroupStrategyOptions defines the strategy type and options used to create the strategy. Deprecated: use FSGroupStrategyOptions from policy API Group instead. - fsGroupStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // HostPortRange defines a range of host ports that will be enabled by a policy for pods to use. It requires both the start and end to be defined. Deprecated: use HostPortRange from policy API Group instead. - hostPortRange:: { - new():: {}, - // max is the end of the range, inclusive. - withMax(max):: self + { max: max }, - // min is the start of the range, inclusive. - withMin(min):: self + { min: min }, - mixin:: {}, - }, - // HTTPIngressPath associates a path regex with a backend. Incoming urls matching the path are forwarded to the backend. - httpIngressPath:: { - new():: {}, - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, (i.e this follows the egrep/unix syntax, not the perl syntax) matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional "path" part of a URL as defined by RFC 3986. Paths must begin with a '/'. If unspecified, the path defaults to a catch all sending traffic to the backend. - withPath(path):: self + { path: path }, - mixin:: { - // Backend defines the referenced service endpoint to which the traffic will be forwarded to. - backend:: { - local __backendMixin(backend) = { backend+: backend }, - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.extensions.v1beta1.ingressBackend, - }, - }, - // HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'. - httpIngressRuleValue:: { - new():: {}, - // A collection of paths that map requests to backends. - withPaths(paths):: self + if std.type(paths) == 'array' then { paths: paths } else { paths: [paths] }, - // A collection of paths that map requests to backends. - withPathsMixin(paths):: self + if std.type(paths) == 'array' then { paths+: paths } else { paths+: [paths] }, - pathsType:: hidden.extensions.v1beta1.httpIngressPath, - mixin:: {}, - }, - // IDRange provides a min/max of an allowed range of IDs. Deprecated: use IDRange from policy API Group instead. - idRange:: { - new():: {}, - // max is the end of the range, inclusive. - withMax(max):: self + { max: max }, - // min is the start of the range, inclusive. - withMin(min):: self + { min: min }, - mixin:: {}, - }, - // IngressBackend describes all endpoints for a given service and port. - ingressBackend:: { - new():: {}, - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + { serviceName: serviceName }, - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + { servicePort: servicePort }, - mixin:: {}, - }, - // IngressList is a collection of Ingress. - ingressList:: { - new():: {}, - // Items is the list of Ingress. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Ingress. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.ingress, - mixin:: {}, - }, - // IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue. - ingressRule:: { - new():: {}, - // Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the "host" part of the URI as defined in the RFC: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. - // 2. The `:` delimiter is not respected because ports are not allowed. - // Currently the port of an Ingress is implicitly :80 for http and - // :443 for https. - // Both these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue. - withHost(host):: self + { host: host }, - mixin:: { - http:: { - local __httpMixin(http) = { http+: http }, - mixinInstance(http):: __httpMixin(http), - // A collection of paths that map requests to backends. - withPaths(paths):: self + if std.type(paths) == 'array' then __httpMixin({ paths: paths }) else __httpMixin({ paths: [paths] }), - // A collection of paths that map requests to backends. - withPathsMixin(paths):: self + if std.type(paths) == 'array' then __httpMixin({ paths+: paths }) else __httpMixin({ paths+: [paths] }), - pathsType:: hidden.extensions.v1beta1.httpIngressPath, - }, - httpType:: hidden.extensions.v1beta1.httpIngressRuleValue, - }, - }, - // IngressSpec describes the Ingress the user wishes to exist. - ingressSpec:: { - new():: {}, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.extensions.v1beta1.ingressRule, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTls(tls):: self + if std.type(tls) == 'array' then { tls: tls } else { tls: [tls] }, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTlsMixin(tls):: self + if std.type(tls) == 'array' then { tls+: tls } else { tls+: [tls] }, - tlsType:: hidden.extensions.v1beta1.ingressTls, - mixin:: { - // A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default. - backend:: { - local __backendMixin(backend) = { backend+: backend }, - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.extensions.v1beta1.ingressBackend, - }, - }, - // IngressStatus describe the current state of the Ingress. - ingressStatus:: { - new():: {}, - mixin:: { - // LoadBalancer contains the current status of the load-balancer. - loadBalancer:: { - local __loadBalancerMixin(loadBalancer) = { loadBalancer+: loadBalancer }, - mixinInstance(loadBalancer):: __loadBalancerMixin(loadBalancer), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngress(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress: ingress }) else __loadBalancerMixin({ ingress: [ingress] }), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress+: ingress }) else __loadBalancerMixin({ ingress+: [ingress] }), - ingressType:: hidden.core.v1.loadBalancerIngress, - }, - loadBalancerType:: hidden.core.v1.loadBalancerStatus, - }, - }, - // IngressTLS describes the transport layer security associated with an Ingress. - ingressTls:: { - new():: {}, - // Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified. - withHosts(hosts):: self + if std.type(hosts) == 'array' then { hosts: hosts } else { hosts: [hosts] }, - // Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified. - withHostsMixin(hosts):: self + if std.type(hosts) == 'array' then { hosts+: hosts } else { hosts+: [hosts] }, - // SecretName is the name of the secret used to terminate SSL traffic on 443. Field is left optional to allow SSL routing based on SNI hostname alone. If the SNI host in a listener conflicts with the "Host" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing. - withSecretName(secretName):: self + { secretName: secretName }, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of IPBlock is deprecated by networking/v1/IPBlock. IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule. - ipBlock:: { - new():: {}, - // CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" - withCidr(cidr):: self + { cidr: cidr }, - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExcept(except):: self + if std.type(except) == 'array' then { except: except } else { except: [except] }, - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExceptMixin(except):: self + if std.type(except) == 'array' then { except+: except } else { except+: [except] }, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicyEgressRule is deprecated by networking/v1/NetworkPolicyEgressRule. NetworkPolicyEgressRule describes a particular set of traffic that is allowed out of pods matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and to. This type is beta-level in 1.8 - networkPolicyEgressRule:: { - new():: {}, - // List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.extensions.v1beta1.networkPolicyPort, - // List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list. - withTo(to):: self + if std.type(to) == 'array' then { to: to } else { to: [to] }, - // List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list. - withToMixin(to):: self + if std.type(to) == 'array' then { to+: to } else { to+: [to] }, - toType:: hidden.extensions.v1beta1.networkPolicyPeer, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicyIngressRule is deprecated by networking/v1/NetworkPolicyIngressRule. This NetworkPolicyIngressRule matches traffic if and only if the traffic matches both ports AND from. - networkPolicyIngressRule:: { - new():: {}, - // List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least on item, this rule allows traffic only if the traffic matches at least one item in the from list. - withFrom(from):: self + if std.type(from) == 'array' then { from: from } else { from: [from] }, - // List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least on item, this rule allows traffic only if the traffic matches at least one item in the from list. - withFromMixin(from):: self + if std.type(from) == 'array' then { from+: from } else { from+: [from] }, - fromType:: hidden.extensions.v1beta1.networkPolicyPeer, - // List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.extensions.v1beta1.networkPolicyPort, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicyList is deprecated by networking/v1/NetworkPolicyList. Network Policy List is a list of NetworkPolicy objects. - networkPolicyList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.networkPolicy, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicyPeer is deprecated by networking/v1/NetworkPolicyPeer. - networkPolicyPeer:: { - new():: {}, - mixin:: { - // IPBlock defines policy on a particular IPBlock. If this field is set then neither of the other fields can be. - ipBlock:: { - local __ipBlockMixin(ipBlock) = { ipBlock+: ipBlock }, - mixinInstance(ipBlock):: __ipBlockMixin(ipBlock), - // CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" - withCidr(cidr):: self + __ipBlockMixin({ cidr: cidr }), - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExcept(except):: self + if std.type(except) == 'array' then __ipBlockMixin({ except: except }) else __ipBlockMixin({ except: [except] }), - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExceptMixin(except):: self + if std.type(except) == 'array' then __ipBlockMixin({ except+: except }) else __ipBlockMixin({ except+: [except] }), - }, - ipBlockType:: hidden.extensions.v1beta1.ipBlock, - // Selects Namespaces using cluster-scoped labels. This field follows standard label selector semantics; if present but empty, it selects all namespaces. - // - // If PodSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects all Pods in the Namespaces selected by NamespaceSelector. - namespaceSelector:: { - local __namespaceSelectorMixin(namespaceSelector) = { namespaceSelector+: namespaceSelector }, - mixinInstance(namespaceSelector):: __namespaceSelectorMixin(namespaceSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions+: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __namespaceSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __namespaceSelectorMixin({ matchLabels+: matchLabels }), - }, - namespaceSelectorType:: hidden.meta.v1.labelSelector, - // This is a label selector which selects Pods. This field follows standard label selector semantics; if present but empty, it selects all pods. - // - // If NamespaceSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects the Pods matching PodSelector in the policy's own Namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = { podSelector+: podSelector }, - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicyPort is deprecated by networking/v1/NetworkPolicyPort. - networkPolicyPort:: { - new():: {}, - // If specified, the port on the given protocol. This can either be a numerical or named port on a pod. If this field is not provided, this matches all port names and numbers. If present, only traffic on the specified protocol AND port will be matched. - withPort(port):: self + { port: port }, - // Optional. The protocol (TCP, UDP, or SCTP) which traffic must match. If not specified, this field defaults to TCP. - withProtocol(protocol):: self + { protocol: protocol }, - mixin:: {}, - }, - // DEPRECATED 1.9 - This group version of NetworkPolicySpec is deprecated by networking/v1/NetworkPolicySpec. - networkPolicySpec:: { - new():: {}, - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgress(egress):: self + if std.type(egress) == 'array' then { egress: egress } else { egress: [egress] }, - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgressMixin(egress):: self + if std.type(egress) == 'array' then { egress+: egress } else { egress+: [egress] }, - egressType:: hidden.extensions.v1beta1.networkPolicyEgressRule, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default). - withIngress(ingress):: self + if std.type(ingress) == 'array' then { ingress: ingress } else { ingress: [ingress] }, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default). - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then { ingress+: ingress } else { ingress+: [ingress] }, - ingressType:: hidden.extensions.v1beta1.networkPolicyIngressRule, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypes(policyTypes):: self + if std.type(policyTypes) == 'array' then { policyTypes: policyTypes } else { policyTypes: [policyTypes] }, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypesMixin(policyTypes):: self + if std.type(policyTypes) == 'array' then { policyTypes+: policyTypes } else { policyTypes+: [policyTypes] }, - mixin:: { - // Selects the pods to which this NetworkPolicy object applies. The array of ingress rules is applied to any pods selected by this field. Multiple network policies can select the same set of pods. In this case, the ingress rules for each are combined additively. This field is NOT optional and follows standard label selector semantics. An empty podSelector matches all pods in this namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = { podSelector+: podSelector }, - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // PodSecurityPolicyList is a list of PodSecurityPolicy objects. Deprecated: use PodSecurityPolicyList from policy API Group instead. - podSecurityPolicyList:: { - new():: {}, - // items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.podSecurityPolicy, - mixin:: {}, - }, - // PodSecurityPolicySpec defines the policy enforced. Deprecated: use PodSecurityPolicySpec from policy API Group instead. - podSecurityPolicySpec:: { - new():: {}, - // allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true. - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + { allowPrivilegeEscalation: allowPrivilegeEscalation }, - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDrivers(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then { allowedCSIDrivers: allowedCsiDrivers } else { allowedCSIDrivers: [allowedCsiDrivers] }, - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDriversMixin(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then { allowedCSIDrivers+: allowedCsiDrivers } else { allowedCSIDrivers+: [allowedCsiDrivers] }, - allowedCSIDriversType:: hidden.extensions.v1beta1.allowedCsiDriver, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilities(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then { allowedCapabilities: allowedCapabilities } else { allowedCapabilities: [allowedCapabilities] }, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilitiesMixin(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then { allowedCapabilities+: allowedCapabilities } else { allowedCapabilities+: [allowedCapabilities] }, - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumes(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then { allowedFlexVolumes: allowedFlexVolumes } else { allowedFlexVolumes: [allowedFlexVolumes] }, - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumesMixin(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then { allowedFlexVolumes+: allowedFlexVolumes } else { allowedFlexVolumes+: [allowedFlexVolumes] }, - allowedFlexVolumesType:: hidden.extensions.v1beta1.allowedFlexVolume, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPaths(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then { allowedHostPaths: allowedHostPaths } else { allowedHostPaths: [allowedHostPaths] }, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPathsMixin(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then { allowedHostPaths+: allowedHostPaths } else { allowedHostPaths+: [allowedHostPaths] }, - allowedHostPathsType:: hidden.extensions.v1beta1.allowedHostPath, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypes(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then { allowedProcMountTypes: allowedProcMountTypes } else { allowedProcMountTypes: [allowedProcMountTypes] }, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypesMixin(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then { allowedProcMountTypes+: allowedProcMountTypes } else { allowedProcMountTypes+: [allowedProcMountTypes] }, - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctls(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then { allowedUnsafeSysctls: allowedUnsafeSysctls } else { allowedUnsafeSysctls: [allowedUnsafeSysctls] }, - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then { allowedUnsafeSysctls+: allowedUnsafeSysctls } else { allowedUnsafeSysctls+: [allowedUnsafeSysctls] }, - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilities(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then { defaultAddCapabilities: defaultAddCapabilities } else { defaultAddCapabilities: [defaultAddCapabilities] }, - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilitiesMixin(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then { defaultAddCapabilities+: defaultAddCapabilities } else { defaultAddCapabilities+: [defaultAddCapabilities] }, - // defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process. - withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation):: self + { defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation }, - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctls(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then { forbiddenSysctls: forbiddenSysctls } else { forbiddenSysctls: [forbiddenSysctls] }, - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctlsMixin(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then { forbiddenSysctls+: forbiddenSysctls } else { forbiddenSysctls+: [forbiddenSysctls] }, - // hostIPC determines if the policy allows the use of HostIPC in the pod spec. - withHostIpc(hostIpc):: self + { hostIPC: hostIpc }, - // hostNetwork determines if the policy allows the use of HostNetwork in the pod spec. - withHostNetwork(hostNetwork):: self + { hostNetwork: hostNetwork }, - // hostPID determines if the policy allows the use of HostPID in the pod spec. - withHostPid(hostPid):: self + { hostPID: hostPid }, - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPorts(hostPorts):: self + if std.type(hostPorts) == 'array' then { hostPorts: hostPorts } else { hostPorts: [hostPorts] }, - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPortsMixin(hostPorts):: self + if std.type(hostPorts) == 'array' then { hostPorts+: hostPorts } else { hostPorts+: [hostPorts] }, - hostPortsType:: hidden.extensions.v1beta1.hostPortRange, - // privileged determines if a pod can request to be run as privileged. - withPrivileged(privileged):: self + { privileged: privileged }, - // readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to. - withReadOnlyRootFilesystem(readOnlyRootFilesystem):: self + { readOnlyRootFilesystem: readOnlyRootFilesystem }, - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilities(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then { requiredDropCapabilities: requiredDropCapabilities } else { requiredDropCapabilities: [requiredDropCapabilities] }, - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilitiesMixin(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then { requiredDropCapabilities+: requiredDropCapabilities } else { requiredDropCapabilities+: [requiredDropCapabilities] }, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumes(volumes):: self + if std.type(volumes) == 'array' then { volumes: volumes } else { volumes: [volumes] }, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then { volumes+: volumes } else { volumes+: [volumes] }, - mixin:: { - // fsGroup is the strategy that will dictate what fs group is used by the SecurityContext. - fsGroup:: { - local __fsGroupMixin(fsGroup) = { fsGroup+: fsGroup }, - mixinInstance(fsGroup):: __fsGroupMixin(fsGroup), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges: ranges }) else __fsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges+: ranges }) else __fsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + __fsGroupMixin({ rule: rule }), - }, - fsGroupType:: hidden.extensions.v1beta1.fsGroupStrategyOptions, - // RunAsGroup is the strategy that will dictate the allowable RunAsGroup values that may be set. If this field is omitted, the pod's RunAsGroup can take any value. This field requires the RunAsGroup feature gate to be enabled. - runAsGroup:: { - local __runAsGroupMixin(runAsGroup) = { runAsGroup+: runAsGroup }, - mixinInstance(runAsGroup):: __runAsGroupMixin(runAsGroup), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges: ranges }) else __runAsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges+: ranges }) else __runAsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + __runAsGroupMixin({ rule: rule }), - }, - runAsGroupType:: hidden.extensions.v1beta1.runAsGroupStrategyOptions, - // runAsUser is the strategy that will dictate the allowable RunAsUser values that may be set. - runAsUser:: { - local __runAsUserMixin(runAsUser) = { runAsUser+: runAsUser }, - mixinInstance(runAsUser):: __runAsUserMixin(runAsUser), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges: ranges }) else __runAsUserMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges+: ranges }) else __runAsUserMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + __runAsUserMixin({ rule: rule }), - }, - runAsUserType:: hidden.extensions.v1beta1.runAsUserStrategyOptions, - // seLinux is the strategy that will dictate the allowable labels that may be set. - seLinux:: { - local __seLinuxMixin(seLinux) = { seLinux+: seLinux }, - mixinInstance(seLinux):: __seLinuxMixin(seLinux), - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + __seLinuxMixin({ rule: rule }), - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __seLinuxMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - seLinuxType:: hidden.extensions.v1beta1.seLinuxStrategyOptions, - // supplementalGroups is the strategy that will dictate what supplemental groups are used by the SecurityContext. - supplementalGroups:: { - local __supplementalGroupsMixin(supplementalGroups) = { supplementalGroups+: supplementalGroups }, - mixinInstance(supplementalGroups):: __supplementalGroupsMixin(supplementalGroups), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges: ranges }) else __supplementalGroupsMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges+: ranges }) else __supplementalGroupsMixin({ ranges+: [ranges] }), - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + __supplementalGroupsMixin({ rule: rule }), - }, - supplementalGroupsType:: hidden.extensions.v1beta1.supplementalGroupsStrategyOptions, - }, - }, - // ReplicaSetCondition describes the state of a replica set at a certain point. - replicaSetCondition:: { - new():: {}, - // The last time the condition transitioned from one status to another. - withLastTransitionTime(lastTransitionTime):: self + { lastTransitionTime: lastTransitionTime }, - // A human readable message indicating details about the transition. - withMessage(message):: self + { message: message }, - // The reason for the condition's last transition. - withReason(reason):: self + { reason: reason }, - // Status of the condition, one of True, False, Unknown. - withStatus(status):: self + { status: status }, - // Type of replica set condition. - withType(type):: self + { type: type }, - mixin:: {}, - }, - // ReplicaSetList is a collection of ReplicaSets. - replicaSetList:: { - new():: {}, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // List of ReplicaSets. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.extensions.v1beta1.replicaSet, - mixin:: {}, - }, - // ReplicaSetSpec is the specification of a ReplicaSet. - replicaSetSpec:: { - new():: {}, - // Minimum number of seconds for which a newly created pod should be ready without any of its container crashing, for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) - withMinReadySeconds(minReadySeconds):: self + { minReadySeconds: minReadySeconds }, - // Replicas is the number of desired replicas. This is a pointer to distinguish between explicit zero and unspecified. Defaults to 1. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: { - // Selector is a label query over pods that should match the replica count. If the selector is empty, it is defaulted to the labels present on the pod template. Label keys and values that must match in order to be controlled by this replica set. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - // Template is the object that describes the pod that will be created if insufficient replicas are detected. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller#pod-template - template:: { - local __templateMixin(template) = { template+: template }, - mixinInstance(template):: __templateMixin(template), - // Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#metadata - metadata:: { - local __metadataMixin(metadata) = __templateMixin({ metadata+: metadata }), - mixinInstance(metadata):: __metadataMixin(metadata), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + __metadataMixin({ annotations: annotations }), - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + __metadataMixin({ annotations+: annotations }), - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + __metadataMixin({ clusterName: clusterName }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers: finalizers }) else __metadataMixin({ finalizers: [finalizers] }), - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then __metadataMixin({ finalizers+: finalizers }) else __metadataMixin({ finalizers+: [finalizers] }), - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + __metadataMixin({ generateName: generateName }), - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = __metadataMixin({ initializers+: initializers }), - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + __metadataMixin({ labels: labels }), - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + __metadataMixin({ labels+: labels }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields: managedFields }) else __metadataMixin({ managedFields: [managedFields] }), - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then __metadataMixin({ managedFields+: managedFields }) else __metadataMixin({ managedFields+: [managedFields] }), - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + __metadataMixin({ name: name }), - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + __metadataMixin({ namespace: namespace }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences: ownerReferences }) else __metadataMixin({ ownerReferences: [ownerReferences] }), - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then __metadataMixin({ ownerReferences+: ownerReferences }) else __metadataMixin({ ownerReferences+: [ownerReferences] }), - ownerReferencesType:: hidden.meta.v1.ownerReference, - }, - metadataType:: hidden.meta.v1.objectMeta, - // Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#spec-and-status - spec:: { - local __specMixin(spec) = __templateMixin({ spec+: spec }), - mixinInstance(spec):: __specMixin(spec), - // Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer. - withActiveDeadlineSeconds(activeDeadlineSeconds):: self + __specMixin({ activeDeadlineSeconds: activeDeadlineSeconds }), - // If specified, the pod's scheduling constraints - affinity:: { - local __affinityMixin(affinity) = __specMixin({ affinity+: affinity }), - mixinInstance(affinity):: __affinityMixin(affinity), - // Describes node affinity scheduling rules for the pod. - nodeAffinity:: { - local __nodeAffinityMixin(nodeAffinity) = __affinityMixin({ nodeAffinity+: nodeAffinity }), - mixinInstance(nodeAffinity):: __nodeAffinityMixin(nodeAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __nodeAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.preferredSchedulingTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node. - requiredDuringSchedulingIgnoredDuringExecution:: { - local __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution) = __nodeAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }), - mixinInstance(requiredDuringSchedulingIgnoredDuringExecution):: __requiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTerms(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms: [nodeSelectorTerms] }), - // Required. A list of node selector terms. The terms are ORed. - withNodeSelectorTermsMixin(nodeSelectorTerms):: self + if std.type(nodeSelectorTerms) == 'array' then __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: nodeSelectorTerms }) else __requiredDuringSchedulingIgnoredDuringExecutionMixin({ nodeSelectorTerms+: [nodeSelectorTerms] }), - nodeSelectorTermsType:: hidden.core.v1.nodeSelectorTerm, - }, - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.nodeSelector, - }, - nodeAffinityType:: hidden.core.v1.nodeAffinity, - // Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s)). - podAffinity:: { - local __podAffinityMixin(podAffinity) = __affinityMixin({ podAffinity+: podAffinity }), - mixinInstance(podAffinity):: __podAffinityMixin(podAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAffinityType:: hidden.core.v1.podAffinity, - // Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s)). - podAntiAffinity:: { - local __podAntiAffinityMixin(podAntiAffinity) = __affinityMixin({ podAntiAffinity+: podAntiAffinity }), - mixinInstance(podAntiAffinity):: __podAntiAffinityMixin(podAntiAffinity), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution: [preferredDuringSchedulingIgnoredDuringExecution] }), - // The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding "weight" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred. - withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution):: self + if std.type(preferredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: preferredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ preferredDuringSchedulingIgnoredDuringExecution+: [preferredDuringSchedulingIgnoredDuringExecution] }), - preferredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.weightedPodAffinityTerm, - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution: [requiredDuringSchedulingIgnoredDuringExecution] }), - // If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied. - withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution):: self + if std.type(requiredDuringSchedulingIgnoredDuringExecution) == 'array' then __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: requiredDuringSchedulingIgnoredDuringExecution }) else __podAntiAffinityMixin({ requiredDuringSchedulingIgnoredDuringExecution+: [requiredDuringSchedulingIgnoredDuringExecution] }), - requiredDuringSchedulingIgnoredDuringExecutionType:: hidden.core.v1.podAffinityTerm, - }, - podAntiAffinityType:: hidden.core.v1.podAntiAffinity, - }, - affinityType:: hidden.core.v1.affinity, - // AutomountServiceAccountToken indicates whether a service account token should be automatically mounted. - withAutomountServiceAccountToken(automountServiceAccountToken):: self + __specMixin({ automountServiceAccountToken: automountServiceAccountToken }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainers(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers: containers }) else __specMixin({ containers: [containers] }), - // List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated. - withContainersMixin(containers):: self + if std.type(containers) == 'array' then __specMixin({ containers+: containers }) else __specMixin({ containers+: [containers] }), - containersType:: hidden.core.v1.container, - // Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy. - dnsConfig:: { - local __dnsConfigMixin(dnsConfig) = __specMixin({ dnsConfig+: dnsConfig }), - mixinInstance(dnsConfig):: __dnsConfigMixin(dnsConfig), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameservers(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers: nameservers }) else __dnsConfigMixin({ nameservers: [nameservers] }), - // A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed. - withNameserversMixin(nameservers):: self + if std.type(nameservers) == 'array' then __dnsConfigMixin({ nameservers+: nameservers }) else __dnsConfigMixin({ nameservers+: [nameservers] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptions(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options: options }) else __dnsConfigMixin({ options: [options] }), - // A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy. - withOptionsMixin(options):: self + if std.type(options) == 'array' then __dnsConfigMixin({ options+: options }) else __dnsConfigMixin({ options+: [options] }), - optionsType:: hidden.core.v1.podDnsConfigOption, - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearches(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches: searches }) else __dnsConfigMixin({ searches: [searches] }), - // A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed. - withSearchesMixin(searches):: self + if std.type(searches) == 'array' then __dnsConfigMixin({ searches+: searches }) else __dnsConfigMixin({ searches+: [searches] }), - }, - dnsConfigType:: hidden.core.v1.podDnsConfig, - // Set DNS policy for the pod. Defaults to "ClusterFirst". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'. - withDnsPolicy(dnsPolicy):: self + __specMixin({ dnsPolicy: dnsPolicy }), - // EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true. - withEnableServiceLinks(enableServiceLinks):: self + __specMixin({ enableServiceLinks: enableServiceLinks }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliases(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases: hostAliases }) else __specMixin({ hostAliases: [hostAliases] }), - // HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods. - withHostAliasesMixin(hostAliases):: self + if std.type(hostAliases) == 'array' then __specMixin({ hostAliases+: hostAliases }) else __specMixin({ hostAliases+: [hostAliases] }), - hostAliasesType:: hidden.core.v1.hostAlias, - // Use the host's ipc namespace. Optional: Default to false. - withHostIpc(hostIpc):: self + __specMixin({ hostIPC: hostIpc }), - // Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false. - withHostNetwork(hostNetwork):: self + __specMixin({ hostNetwork: hostNetwork }), - // Use the host's pid namespace. Optional: Default to false. - withHostPid(hostPid):: self + __specMixin({ hostPID: hostPid }), - // Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value. - withHostname(hostname):: self + __specMixin({ hostname: hostname }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecrets(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets: imagePullSecrets }) else __specMixin({ imagePullSecrets: [imagePullSecrets] }), - // ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. For example, in the case of docker, only DockerConfig type secrets are honored. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod - withImagePullSecretsMixin(imagePullSecrets):: self + if std.type(imagePullSecrets) == 'array' then __specMixin({ imagePullSecrets+: imagePullSecrets }) else __specMixin({ imagePullSecrets+: [imagePullSecrets] }), - imagePullSecretsType:: hidden.core.v1.localObjectReference, - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainers(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers: initContainers }) else __specMixin({ initContainers: [initContainers] }), - // List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, or Liveness probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - withInitContainersMixin(initContainers):: self + if std.type(initContainers) == 'array' then __specMixin({ initContainers+: initContainers }) else __specMixin({ initContainers+: [initContainers] }), - initContainersType:: hidden.core.v1.container, - // NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements. - withNodeName(nodeName):: self + __specMixin({ nodeName: nodeName }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelector(nodeSelector):: self + __specMixin({ nodeSelector: nodeSelector }), - // NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ - withNodeSelectorMixin(nodeSelector):: self + __specMixin({ nodeSelector+: nodeSelector }), - // The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority. - withPriority(priority):: self + __specMixin({ priority: priority }), - // If specified, indicates the pod's priority. "system-node-critical" and "system-cluster-critical" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default. - withPriorityClassName(priorityClassName):: self + __specMixin({ priorityClassName: priorityClassName }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGates(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates: readinessGates }) else __specMixin({ readinessGates: [readinessGates] }), - // If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to "True" More info: https://git.k8s.io/enhancements/keps/sig-network/0007-pod-ready%2B%2B.md - withReadinessGatesMixin(readinessGates):: self + if std.type(readinessGates) == 'array' then __specMixin({ readinessGates+: readinessGates }) else __specMixin({ readinessGates+: [readinessGates] }), - readinessGatesType:: hidden.core.v1.podReadinessGate, - // Restart policy for all containers within the pod. One of Always, OnFailure, Never. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy - withRestartPolicy(restartPolicy):: self + __specMixin({ restartPolicy: restartPolicy }), - // RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the "legacy" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/runtime-class.md This is an alpha feature and may change in the future. - withRuntimeClassName(runtimeClassName):: self + __specMixin({ runtimeClassName: runtimeClassName }), - // If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler. - withSchedulerName(schedulerName):: self + __specMixin({ schedulerName: schedulerName }), - // SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field. - securityContext:: { - local __securityContextMixin(securityContext) = __specMixin({ securityContext+: securityContext }), - mixinInstance(securityContext):: __securityContextMixin(securityContext), - // A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: - // - // 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- - // - // If unset, the Kubelet will not modify the ownership and permissions of any volume. - withFsGroup(fsGroup):: self + __securityContextMixin({ fsGroup: fsGroup }), - // The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsGroup(runAsGroup):: self + __securityContextMixin({ runAsGroup: runAsGroup }), - // Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. - withRunAsNonRoot(runAsNonRoot):: self + __securityContextMixin({ runAsNonRoot: runAsNonRoot }), - // The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - withRunAsUser(runAsUser):: self + __securityContextMixin({ runAsUser: runAsUser }), - // The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __securityContextMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroups(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups: supplementalGroups }) else __securityContextMixin({ supplementalGroups: [supplementalGroups] }), - // A list of groups applied to the first process run in each container, in addition to the container's primary GID. If unspecified, no groups will be added to any container. - withSupplementalGroupsMixin(supplementalGroups):: self + if std.type(supplementalGroups) == 'array' then __securityContextMixin({ supplementalGroups+: supplementalGroups }) else __securityContextMixin({ supplementalGroups+: [supplementalGroups] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctls(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls: sysctls }) else __securityContextMixin({ sysctls: [sysctls] }), - // Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. - withSysctlsMixin(sysctls):: self + if std.type(sysctls) == 'array' then __securityContextMixin({ sysctls+: sysctls }) else __securityContextMixin({ sysctls+: [sysctls] }), - sysctlsType:: hidden.core.v1.sysctl, - }, - securityContextType:: hidden.core.v1.podSecurityContext, - // DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead. - withServiceAccount(serviceAccount):: self + __specMixin({ serviceAccount: serviceAccount }), - // ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/ - withServiceAccountName(serviceAccountName):: self + __specMixin({ serviceAccountName: serviceAccountName }), - // Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false. This field is beta-level and may be disabled with the PodShareProcessNamespace feature. - withShareProcessNamespace(shareProcessNamespace):: self + __specMixin({ shareProcessNamespace: shareProcessNamespace }), - // If specified, the fully qualified Pod hostname will be "...svc.". If not specified, the pod will not have a domainname at all. - withSubdomain(subdomain):: self + __specMixin({ subdomain: subdomain }), - // Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds. - withTerminationGracePeriodSeconds(terminationGracePeriodSeconds):: self + __specMixin({ terminationGracePeriodSeconds: terminationGracePeriodSeconds }), - // If specified, the pod's tolerations. - withTolerations(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations: tolerations }) else __specMixin({ tolerations: [tolerations] }), - // If specified, the pod's tolerations. - withTolerationsMixin(tolerations):: self + if std.type(tolerations) == 'array' then __specMixin({ tolerations+: tolerations }) else __specMixin({ tolerations+: [tolerations] }), - tolerationsType:: hidden.core.v1.toleration, - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumes(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes: volumes }) else __specMixin({ volumes: [volumes] }), - // List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then __specMixin({ volumes+: volumes }) else __specMixin({ volumes+: [volumes] }), - volumesType:: hidden.core.v1.volume, - }, - specType:: hidden.core.v1.podSpec, - }, - templateType:: hidden.core.v1.podTemplateSpec, - }, - }, - // ReplicaSetStatus represents the current status of a ReplicaSet. - replicaSetStatus:: { - new():: {}, - // The number of available replicas (ready for at least minReadySeconds) for this replica set. - withAvailableReplicas(availableReplicas):: self + { availableReplicas: availableReplicas }, - // Represents the latest available observations of a replica set's current state. - withConditions(conditions):: self + if std.type(conditions) == 'array' then { conditions: conditions } else { conditions: [conditions] }, - // Represents the latest available observations of a replica set's current state. - withConditionsMixin(conditions):: self + if std.type(conditions) == 'array' then { conditions+: conditions } else { conditions+: [conditions] }, - conditionsType:: hidden.extensions.v1beta1.replicaSetCondition, - // The number of pods that have labels matching the labels of the pod template of the replicaset. - withFullyLabeledReplicas(fullyLabeledReplicas):: self + { fullyLabeledReplicas: fullyLabeledReplicas }, - // ObservedGeneration reflects the generation of the most recently observed ReplicaSet. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - // The number of ready replicas for this replica set. - withReadyReplicas(readyReplicas):: self + { readyReplicas: readyReplicas }, - // Replicas is the most recently oberved number of replicas. More info: https://kubernetes.io/docs/concepts/workloads/controllers/replicationcontroller/#what-is-a-replicationcontroller - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // DEPRECATED. - rollbackConfig:: { - new():: {}, - // The revision to rollback to. If set to 0, rollback to the last revision. - withRevision(revision):: self + { revision: revision }, - mixin:: {}, - }, - // Spec to control the desired behavior of daemon set rolling update. - rollingUpdateDaemonSet:: { - new():: {}, - // The maximum number of DaemonSet pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of total number of DaemonSet pods at the start of the update (ex: 10%). Absolute number is calculated from percentage by rounding up. This cannot be 0. Default value is 1. Example: when this is set to 30%, at most 30% of the total number of nodes that should be running the daemon pod (i.e. status.desiredNumberScheduled) can have their pods stopped for an update at any given time. The update starts by stopping at most 30% of those DaemonSet pods and then brings up new DaemonSet pods in their place. Once the new pods are available, it then proceeds onto other DaemonSet pods, thus ensuring that at least 70% of original number of DaemonSet pods are available at all times during the update. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // Spec to control the desired behavior of rolling update. - rollingUpdateDeployment:: { - new():: {}, - // The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. By default, a value of 1 is used. Example: when this is set to 30%, the new RC can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new RC can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods. - withMaxSurge(maxSurge):: self + { maxSurge: maxSurge }, - // The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. By default, a fixed value of 1 is used. Example: when this is set to 30%, the old RC can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old RC can be scaled down further, followed by scaling up the new RC, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods. - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - mixin:: {}, - }, - // RunAsGroupStrategyOptions defines the strategy type and any options used to create the strategy. Deprecated: use RunAsGroupStrategyOptions from policy API Group instead. - runAsGroupStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy. Deprecated: use RunAsUserStrategyOptions from policy API Group instead. - runAsUserStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // describes the attributes of a scale subresource - scaleSpec:: { - new():: {}, - // desired number of instances for the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - mixin:: {}, - }, - // represents the current status of a scale subresource. - scaleStatus:: { - new():: {}, - // actual number of observed instances of the scaled object. - withReplicas(replicas):: self + { replicas: replicas }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelector(selector):: self + { selector: selector }, - // label query over pods that should match the replicas count. More info: http://kubernetes.io/docs/user-guide/labels#label-selectors - withSelectorMixin(selector):: self + { selector+: selector }, - // label selector for pods that should match the replicas count. This is a serializated version of both map-based and more expressive set-based selectors. This is done to avoid introspection in the clients. The string will be in the same format as the query-param syntax. If the target type only supports map-based selectors, both this field and map-based selector field are populated. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/#label-selectors - withTargetSelector(targetSelector):: self + { targetSelector: targetSelector }, - mixin:: {}, - }, - // SELinuxStrategyOptions defines the strategy type and any options used to create the strategy. Deprecated: use SELinuxStrategyOptions from policy API Group instead. - seLinuxStrategyOptions:: { - new():: {}, - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: { - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = { seLinuxOptions+: seLinuxOptions }, - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - }, - // SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy. Deprecated: use SupplementalGroupsStrategyOptions from policy API Group instead. - supplementalGroupsStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.extensions.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - }, - }, - meta:: { - v1:: { - local apiVersion = { apiVersion: 'meta/v1' }, - // APIGroup contains the name, the supported versions, and the preferred version of a group. - apiGroup:: { - new():: {}, - // name is the name of the group. - withName(name):: self + { name: name }, - // a map of client CIDR to server address that is serving this group. This is to help clients reach servers in the most network-efficient way possible. Clients can use the appropriate server address as per the CIDR that they match. In case of multiple matches, clients should use the longest matching CIDR. The server returns only those CIDRs that it thinks that the client can match. For example: the master will return an internal IP CIDR only, if the client reaches the server using an internal IP. Server looks at X-Forwarded-For header or X-Real-Ip header or request.RemoteAddr (in that order) to get the client IP. - withServerAddressByClientCidrs(serverAddressByClientCidrs):: self + if std.type(serverAddressByClientCidrs) == 'array' then { serverAddressByClientCIDRs: serverAddressByClientCidrs } else { serverAddressByClientCIDRs: [serverAddressByClientCidrs] }, - // a map of client CIDR to server address that is serving this group. This is to help clients reach servers in the most network-efficient way possible. Clients can use the appropriate server address as per the CIDR that they match. In case of multiple matches, clients should use the longest matching CIDR. The server returns only those CIDRs that it thinks that the client can match. For example: the master will return an internal IP CIDR only, if the client reaches the server using an internal IP. Server looks at X-Forwarded-For header or X-Real-Ip header or request.RemoteAddr (in that order) to get the client IP. - withServerAddressByClientCidrsMixin(serverAddressByClientCidrs):: self + if std.type(serverAddressByClientCidrs) == 'array' then { serverAddressByClientCIDRs+: serverAddressByClientCidrs } else { serverAddressByClientCIDRs+: [serverAddressByClientCidrs] }, - serverAddressByClientCIDRsType:: hidden.meta.v1.serverAddressByClientCidr, - // versions are the versions supported in this group. - withVersions(versions):: self + if std.type(versions) == 'array' then { versions: versions } else { versions: [versions] }, - // versions are the versions supported in this group. - withVersionsMixin(versions):: self + if std.type(versions) == 'array' then { versions+: versions } else { versions+: [versions] }, - versionsType:: hidden.meta.v1.groupVersionForDiscovery, - mixin:: { - // preferredVersion is the version preferred by the API server, which probably is the storage version. - preferredVersion:: { - local __preferredVersionMixin(preferredVersion) = { preferredVersion+: preferredVersion }, - mixinInstance(preferredVersion):: __preferredVersionMixin(preferredVersion), - // groupVersion specifies the API group and version in the form "group/version" - withGroupVersion(groupVersion):: self + __preferredVersionMixin({ groupVersion: groupVersion }), - // version specifies the version in the form of "version". This is to save the clients the trouble of splitting the GroupVersion. - withVersion(version):: self + __preferredVersionMixin({ version: version }), - }, - preferredVersionType:: hidden.meta.v1.groupVersionForDiscovery, - }, - }, - // APIGroupList is a list of APIGroup, to allow clients to discover the API at /apis. - apiGroupList:: { - new():: {}, - // groups is a list of APIGroup. - withGroups(groups):: self + if std.type(groups) == 'array' then { groups: groups } else { groups: [groups] }, - // groups is a list of APIGroup. - withGroupsMixin(groups):: self + if std.type(groups) == 'array' then { groups+: groups } else { groups+: [groups] }, - groupsType:: hidden.meta.v1.apiGroup, - mixin:: {}, - }, - // APIResource specifies the name of a resource and whether it is namespaced. - apiResource:: { - new():: {}, - // categories is a list of the grouped resources this resource belongs to (e.g. 'all') - withCategories(categories):: self + if std.type(categories) == 'array' then { categories: categories } else { categories: [categories] }, - // categories is a list of the grouped resources this resource belongs to (e.g. 'all') - withCategoriesMixin(categories):: self + if std.type(categories) == 'array' then { categories+: categories } else { categories+: [categories] }, - // group is the preferred group of the resource. Empty implies the group of the containing resource list. For subresources, this may have a different value, for example: Scale". - withGroup(group):: self + { group: group }, - // kind is the kind for the resource (e.g. 'Foo' is the kind for a resource 'foo') - withKind(kind):: self + { kind: kind }, - // name is the plural name of the resource. - withName(name):: self + { name: name }, - // namespaced indicates if a resource is namespaced or not. - withNamespaced(namespaced):: self + { namespaced: namespaced }, - // shortNames is a list of suggested short names of the resource. - withShortNames(shortNames):: self + if std.type(shortNames) == 'array' then { shortNames: shortNames } else { shortNames: [shortNames] }, - // shortNames is a list of suggested short names of the resource. - withShortNamesMixin(shortNames):: self + if std.type(shortNames) == 'array' then { shortNames+: shortNames } else { shortNames+: [shortNames] }, - // singularName is the singular name of the resource. This allows clients to handle plural and singular opaquely. The singularName is more correct for reporting status on a single item and both singular and plural are allowed from the kubectl CLI interface. - withSingularName(singularName):: self + { singularName: singularName }, - // The hash value of the storage version, the version this resource is converted to when written to the data store. Value must be treated as opaque by clients. Only equality comparison on the value is valid. This is an alpha feature and may change or be removed in the future. The field is populated by the apiserver only if the StorageVersionHash feature gate is enabled. This field will remain optional even if it graduates. - withStorageVersionHash(storageVersionHash):: self + { storageVersionHash: storageVersionHash }, - // verbs is a list of supported kube verbs (this includes get, list, watch, create, update, patch, delete, deletecollection, and proxy) - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // verbs is a list of supported kube verbs (this includes get, list, watch, create, update, patch, delete, deletecollection, and proxy) - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - // version is the preferred version of the resource. Empty implies the version of the containing resource list For subresources, this may have a different value, for example: v1 (while inside a v1beta1 version of the core resource's group)". - withVersion(version):: self + { version: version }, - mixin:: {}, - }, - // APIResourceList is a list of APIResource, it is used to expose the name of the resources supported in a specific group and version, and if the resource is namespaced. - apiResourceList:: { - new():: {}, - // groupVersion is the group and version this APIResourceList is for. - withGroupVersion(groupVersion):: self + { groupVersion: groupVersion }, - // resources contains the name of the resources and if they are namespaced. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // resources contains the name of the resources and if they are namespaced. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - resourcesType:: hidden.meta.v1.apiResource, - mixin:: {}, - }, - // APIVersions lists the versions that are available, to allow clients to discover the API at /api, which is the root path of the legacy v1 API. - apiVersions:: { - new():: {}, - // a map of client CIDR to server address that is serving this group. This is to help clients reach servers in the most network-efficient way possible. Clients can use the appropriate server address as per the CIDR that they match. In case of multiple matches, clients should use the longest matching CIDR. The server returns only those CIDRs that it thinks that the client can match. For example: the master will return an internal IP CIDR only, if the client reaches the server using an internal IP. Server looks at X-Forwarded-For header or X-Real-Ip header or request.RemoteAddr (in that order) to get the client IP. - withServerAddressByClientCidrs(serverAddressByClientCidrs):: self + if std.type(serverAddressByClientCidrs) == 'array' then { serverAddressByClientCIDRs: serverAddressByClientCidrs } else { serverAddressByClientCIDRs: [serverAddressByClientCidrs] }, - // a map of client CIDR to server address that is serving this group. This is to help clients reach servers in the most network-efficient way possible. Clients can use the appropriate server address as per the CIDR that they match. In case of multiple matches, clients should use the longest matching CIDR. The server returns only those CIDRs that it thinks that the client can match. For example: the master will return an internal IP CIDR only, if the client reaches the server using an internal IP. Server looks at X-Forwarded-For header or X-Real-Ip header or request.RemoteAddr (in that order) to get the client IP. - withServerAddressByClientCidrsMixin(serverAddressByClientCidrs):: self + if std.type(serverAddressByClientCidrs) == 'array' then { serverAddressByClientCIDRs+: serverAddressByClientCidrs } else { serverAddressByClientCIDRs+: [serverAddressByClientCidrs] }, - serverAddressByClientCIDRsType:: hidden.meta.v1.serverAddressByClientCidr, - // versions are the api versions that are available. - withVersions(versions):: self + if std.type(versions) == 'array' then { versions: versions } else { versions: [versions] }, - // versions are the api versions that are available. - withVersionsMixin(versions):: self + if std.type(versions) == 'array' then { versions+: versions } else { versions+: [versions] }, - mixin:: {}, - }, - // DeleteOptions may be provided when deleting an API object. - deleteOptions:: { - new():: {}, - // When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed - withDryRun(dryRun):: self + if std.type(dryRun) == 'array' then { dryRun: dryRun } else { dryRun: [dryRun] }, - // When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed - withDryRunMixin(dryRun):: self + if std.type(dryRun) == 'array' then { dryRun+: dryRun } else { dryRun+: [dryRun] }, - // The duration in seconds before the object should be deleted. Value must be non-negative integer. The value zero indicates delete immediately. If this value is nil, the default grace period for the specified type will be used. Defaults to a per object value if not specified. zero means delete immediately. - withGracePeriodSeconds(gracePeriodSeconds):: self + { gracePeriodSeconds: gracePeriodSeconds }, - // Deprecated: please use the PropagationPolicy, this field will be deprecated in 1.7. Should the dependent objects be orphaned. If true/false, the "orphan" finalizer will be added to/removed from the object's finalizers list. Either this field or PropagationPolicy may be set, but not both. - withOrphanDependents(orphanDependents):: self + { orphanDependents: orphanDependents }, - // Whether and how garbage collection will be performed. Either this field or OrphanDependents may be set, but not both. The default policy is decided by the existing finalizer set in the metadata.finalizers and the resource-specific default policy. Acceptable values are: 'Orphan' - orphan the dependents; 'Background' - allow the garbage collector to delete the dependents in the background; 'Foreground' - a cascading policy that deletes all dependents in the foreground. - withPropagationPolicy(propagationPolicy):: self + { propagationPolicy: propagationPolicy }, - mixin:: { - // Must be fulfilled before a deletion is carried out. If not possible, a 409 Conflict status will be returned. - preconditions:: { - local __preconditionsMixin(preconditions) = { preconditions+: preconditions }, - mixinInstance(preconditions):: __preconditionsMixin(preconditions), - // Specifies the target ResourceVersion - withResourceVersion(resourceVersion):: self + __preconditionsMixin({ resourceVersion: resourceVersion }), - // Specifies the target UID. - withUid(uid):: self + __preconditionsMixin({ uid: uid }), - }, - preconditionsType:: hidden.meta.v1.preconditions, - }, - }, - // Fields stores a set of fields in a data structure like a Trie. To understand how this is used, see: https://github.com/kubernetes-sigs/structured-merge-diff - fields:: { - new():: {}, - mixin:: {}, - }, - // GroupVersion contains the "group/version" and "version" string of a version. It is made a struct to keep extensibility. - groupVersionForDiscovery:: { - new():: {}, - // groupVersion specifies the API group and version in the form "group/version" - withGroupVersion(groupVersion):: self + { groupVersion: groupVersion }, - // version specifies the version in the form of "version". This is to save the clients the trouble of splitting the GroupVersion. - withVersion(version):: self + { version: version }, - mixin:: {}, - }, - // Initializer is information about an initializer that has not yet completed. - initializer:: { - new():: {}, - // name of the process that is responsible for initializing this object. - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Initializers tracks the progress of initialization. - initializers:: { - new():: {}, - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then { pending: pending } else { pending: [pending] }, - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then { pending+: pending } else { pending+: [pending] }, - pendingType:: hidden.meta.v1.initializer, - mixin:: {}, - }, - // A label selector is a label query over a set of resources. The result of matchLabels and matchExpressions are ANDed. An empty label selector matches all objects. A null label selector matches no objects. - labelSelector:: { - new():: {}, - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions: matchExpressions } else { matchExpressions: [matchExpressions] }, - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then { matchExpressions+: matchExpressions } else { matchExpressions+: [matchExpressions] }, - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + { matchLabels: matchLabels }, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + { matchLabels+: matchLabels }, - mixin:: {}, - }, - // A label selector requirement is a selector that contains values, a key, and an operator that relates the key and values. - labelSelectorRequirement:: { - new():: {}, - // key is the label key that the selector applies to. - withKey(key):: self + { key: key }, - // operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist. - withOperator(operator):: self + { operator: operator }, - // values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. - withValues(values):: self + if std.type(values) == 'array' then { values: values } else { values: [values] }, - // values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch. - withValuesMixin(values):: self + if std.type(values) == 'array' then { values+: values } else { values+: [values] }, - mixin:: {}, - }, - // ListMeta describes metadata that synthetic resources must have, including lists and various status objects. A resource may have only one of {ObjectMeta, ListMeta}. - listMeta:: { - new():: {}, - // continue may be set if the user set a limit on the number of items returned, and indicates that the server has more data available. The value is opaque and may be used to issue another request to the endpoint that served this list to retrieve the next set of available objects. Continuing a consistent list may not be possible if the server configuration has changed or more than a few minutes have passed. The resourceVersion field returned when using this continue value will be identical to the value in the first response, unless you have received this token from an error message. - withContinue(continue):: self + { continue: continue }, - mixin:: {}, - }, - // ManagedFieldsEntry is a workflow-id, a FieldSet and the group version of the resource that the fieldset applies to. - managedFieldsEntry:: { - new():: {}, - // Manager is an identifier of the workflow managing these fields. - withManager(manager):: self + { manager: manager }, - // Operation is the type of operation which lead to this ManagedFieldsEntry being created. The only valid values for this field are 'Apply' and 'Update'. - withOperation(operation):: self + { operation: operation }, - // Time is timestamp of when these fields were set. It should always be empty if Operation is 'Apply' - withTime(time):: self + { time: time }, - mixin:: { - // Fields identifies a set of fields. - fields:: { - local __fieldsMixin(fields) = { fields+: fields }, - mixinInstance(fields):: __fieldsMixin(fields), - }, - fieldsType:: hidden.meta.v1.fields, - }, - }, - // MicroTime is version of Time with microsecond level precision. - microTime:: { - new():: {}, - mixin:: {}, - }, - // ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create. - objectMeta:: { - new():: {}, - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotations(annotations):: self + { annotations: annotations }, - // Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations - withAnnotationsMixin(annotations):: self + { annotations+: annotations }, - // The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request. - withClusterName(clusterName):: self + { clusterName: clusterName }, - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizers(finalizers):: self + if std.type(finalizers) == 'array' then { finalizers: finalizers } else { finalizers: [finalizers] }, - // Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. - withFinalizersMixin(finalizers):: self + if std.type(finalizers) == 'array' then { finalizers+: finalizers } else { finalizers+: [finalizers] }, - // GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server. - // - // If this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header). - // - // Applied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#idempotency - withGenerateName(generateName):: self + { generateName: generateName }, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabels(labels):: self + { labels: labels }, - // Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels - withLabelsMixin(labels):: self + { labels+: labels }, - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFields(managedFields):: self + if std.type(managedFields) == 'array' then { managedFields: managedFields } else { managedFields: [managedFields] }, - // ManagedFields maps workflow-id and version to the set of fields that are managed by that workflow. This is mostly for internal housekeeping, and users typically shouldn't need to set or understand this field. A workflow can be the user's name, a controller's name, or the name of a specific apply path like "ci-cd". The set of fields is always in the version that the workflow used when modifying the object. - // - // This field is alpha and can be changed or removed without notice. - withManagedFieldsMixin(managedFields):: self + if std.type(managedFields) == 'array' then { managedFields+: managedFields } else { managedFields+: [managedFields] }, - managedFieldsType:: hidden.meta.v1.managedFieldsEntry, - // Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + { name: name }, - // Namespace defines the space within each name must be unique. An empty namespace is equivalent to the "default" namespace, but "default" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty. - // - // Must be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces - withNamespace(namespace):: self + { namespace: namespace }, - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferences(ownerReferences):: self + if std.type(ownerReferences) == 'array' then { ownerReferences: ownerReferences } else { ownerReferences: [ownerReferences] }, - // List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller. - withOwnerReferencesMixin(ownerReferences):: self + if std.type(ownerReferences) == 'array' then { ownerReferences+: ownerReferences } else { ownerReferences+: [ownerReferences] }, - ownerReferencesType:: hidden.meta.v1.ownerReference, - mixin:: { - // An initializer is a controller which enforces some system invariant at object creation time. This field is a list of initializers that have not yet acted on this object. If nil or empty, this object has been completely initialized. Otherwise, the object is considered uninitialized and is hidden (in list/watch and get calls) from clients that haven't explicitly asked to observe uninitialized objects. - // - // When an object is created, the system will populate this list with the current set of initializers. Only privileged users may set or modify this list. Once it is empty, it may not be modified further by any user. - // - // DEPRECATED - initializers are an alpha field and will be removed in v1.15. - initializers:: { - local __initializersMixin(initializers) = { initializers+: initializers }, - mixinInstance(initializers):: __initializersMixin(initializers), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPending(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending: pending }) else __initializersMixin({ pending: [pending] }), - // Pending is a list of initializers that must execute in order before this object is visible. When the last pending initializer is removed, and no failing result is set, the initializers struct will be set to nil and the object is considered as initialized and visible to all clients. - withPendingMixin(pending):: self + if std.type(pending) == 'array' then __initializersMixin({ pending+: pending }) else __initializersMixin({ pending+: [pending] }), - pendingType:: hidden.meta.v1.initializer, - }, - initializersType:: hidden.meta.v1.initializers, - }, - }, - // OwnerReference contains enough information to let you identify an owning object. An owning object must be in the same namespace as the dependent, or be cluster-scoped, so there is no namespace field. - ownerReference:: { - new():: {}, - // API version of the referent. - withApiVersion(apiVersion):: self + { apiVersion: apiVersion }, - // If true, AND if the owner has the "foregroundDeletion" finalizer, then the owner cannot be deleted from the key-value store until this reference is removed. Defaults to false. To set this field, a user needs "delete" permission of the owner, otherwise 422 (Unprocessable Entity) will be returned. - withBlockOwnerDeletion(blockOwnerDeletion):: self + { blockOwnerDeletion: blockOwnerDeletion }, - // If true, this reference points to the managing controller. - withController(controller):: self + { controller: controller }, - // Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/api-conventions.md#types-kinds - withKind(kind):: self + { kind: kind }, - // Name of the referent. More info: http://kubernetes.io/docs/user-guide/identifiers#names - withName(name):: self + { name: name }, - // UID of the referent. More info: http://kubernetes.io/docs/user-guide/identifiers#uids - withUid(uid):: self + { uid: uid }, - mixin:: {}, - }, - // Preconditions must be fulfilled before an operation (update, delete, etc.) is carried out. - preconditions:: { - new():: {}, - // Specifies the target ResourceVersion - withResourceVersion(resourceVersion):: self + { resourceVersion: resourceVersion }, - // Specifies the target UID. - withUid(uid):: self + { uid: uid }, - mixin:: {}, - }, - // ServerAddressByClientCIDR helps the client to determine the server address that they should use, depending on the clientCIDR that they match. - serverAddressByClientCidr:: { - new():: {}, - // The CIDR with which clients can match their IP to figure out the server address that they should use. - withClientCidr(clientCidr):: self + { clientCIDR: clientCidr }, - // Address of this server, suitable for a client that matches the above CIDR. This can be a hostname, hostname:port, IP or IP:port. - withServerAddress(serverAddress):: self + { serverAddress: serverAddress }, - mixin:: {}, - }, - // Status is a return value for calls that don't return other objects. - status:: { - new():: {}, - // Suggested HTTP return code for this status, 0 if not set. - withCode(code):: self + { code: code }, - // A human-readable description of the status of this operation. - withMessage(message):: self + { message: message }, - // A machine-readable description of why this operation is in the "Failure" status. If this value is empty there is no information available. A Reason clarifies an HTTP status code but does not override it. - withReason(reason):: self + { reason: reason }, - mixin:: { - // Extended data associated with the reason. Each reason may define its own extended details. This field is optional and the data returned is not guaranteed to conform to any schema except that defined by the reason type. - details:: { - local __detailsMixin(details) = { details+: details }, - mixinInstance(details):: __detailsMixin(details), - // The Causes array includes more details associated with the StatusReason failure. Not all StatusReasons may provide detailed causes. - withCauses(causes):: self + if std.type(causes) == 'array' then __detailsMixin({ causes: causes }) else __detailsMixin({ causes: [causes] }), - // The Causes array includes more details associated with the StatusReason failure. Not all StatusReasons may provide detailed causes. - withCausesMixin(causes):: self + if std.type(causes) == 'array' then __detailsMixin({ causes+: causes }) else __detailsMixin({ causes+: [causes] }), - causesType:: hidden.meta.v1.statusCause, - // The group attribute of the resource associated with the status StatusReason. - withGroup(group):: self + __detailsMixin({ group: group }), - // The name attribute of the resource associated with the status StatusReason (when there is a single name which can be described). - withName(name):: self + __detailsMixin({ name: name }), - // If specified, the time in seconds before the operation should be retried. Some errors may indicate the client must take an alternate action - for those errors this field may indicate how long to wait before taking the alternate action. - withRetryAfterSeconds(retryAfterSeconds):: self + __detailsMixin({ retryAfterSeconds: retryAfterSeconds }), - // UID of the resource. (when there is a single resource which can be described). More info: http://kubernetes.io/docs/user-guide/identifiers#uids - withUid(uid):: self + __detailsMixin({ uid: uid }), - }, - detailsType:: hidden.meta.v1.statusDetails, - }, - }, - // StatusCause provides more information about an api.Status failure, including cases when multiple errors are encountered. - statusCause:: { - new():: {}, - // The field of the resource that has caused this error, as named by its JSON serialization. May include dot and postfix notation for nested attributes. Arrays are zero-indexed. Fields may appear more than once in an array of causes due to fields having multiple errors. Optional. - // - // Examples: - // "name" - the field "name" on the current resource - // "items[0].name" - the field "name" on the first array entry in "items" - withField(field):: self + { field: field }, - // A human-readable description of the cause of the error. This field may be presented as-is to a reader. - withMessage(message):: self + { message: message }, - // A machine-readable description of the cause of the error. If this value is empty there is no information available. - withReason(reason):: self + { reason: reason }, - mixin:: {}, - }, - // StatusDetails is a set of additional properties that MAY be set by the server to provide additional information about a response. The Reason field of a Status object defines what attributes will be set. Clients must ignore fields that do not match the defined type of each attribute, and should assume that any attribute may be empty, invalid, or under defined. - statusDetails:: { - new():: {}, - // The Causes array includes more details associated with the StatusReason failure. Not all StatusReasons may provide detailed causes. - withCauses(causes):: self + if std.type(causes) == 'array' then { causes: causes } else { causes: [causes] }, - // The Causes array includes more details associated with the StatusReason failure. Not all StatusReasons may provide detailed causes. - withCausesMixin(causes):: self + if std.type(causes) == 'array' then { causes+: causes } else { causes+: [causes] }, - causesType:: hidden.meta.v1.statusCause, - // The group attribute of the resource associated with the status StatusReason. - withGroup(group):: self + { group: group }, - // The name attribute of the resource associated with the status StatusReason (when there is a single name which can be described). - withName(name):: self + { name: name }, - // If specified, the time in seconds before the operation should be retried. Some errors may indicate the client must take an alternate action - for those errors this field may indicate how long to wait before taking the alternate action. - withRetryAfterSeconds(retryAfterSeconds):: self + { retryAfterSeconds: retryAfterSeconds }, - // UID of the resource. (when there is a single resource which can be described). More info: http://kubernetes.io/docs/user-guide/identifiers#uids - withUid(uid):: self + { uid: uid }, - mixin:: {}, - }, - // Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers. - time:: { - new():: {}, - mixin:: {}, - }, - // Event represents a single event to a watched resource. - watchEvent:: { - new():: {}, - withType(type):: self + { type: type }, - mixin:: { - // Object is: - // * If Type is Added or Modified: the new state of the object. - // * If Type is Deleted: the state of the object immediately before deletion. - // * If Type is Error: *Status is recommended; other types may make sense - // depending on context. - object:: { - local __objectMixin(object) = { object+: object }, - mixinInstance(object):: __objectMixin(object), - // Raw is the underlying serialization of this object. - withRaw(raw):: self + __objectMixin({ Raw: raw }), - }, - }, - }, - }, - }, - networking:: { - v1:: { - local apiVersion = { apiVersion: 'networking/v1' }, - // IPBlock describes a particular CIDR (Ex. "192.168.1.1/24") that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The except entry describes CIDRs that should not be included within this rule. - ipBlock:: { - new():: {}, - // CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" - withCidr(cidr):: self + { cidr: cidr }, - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExcept(except):: self + if std.type(except) == 'array' then { except: except } else { except: [except] }, - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExceptMixin(except):: self + if std.type(except) == 'array' then { except+: except } else { except+: [except] }, - mixin:: {}, - }, - // NetworkPolicyEgressRule describes a particular set of traffic that is allowed out of pods matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and to. This type is beta-level in 1.8 - networkPolicyEgressRule:: { - new():: {}, - // List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // List of destination ports for outgoing traffic. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.networking.v1.networkPolicyPort, - // List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list. - withTo(to):: self + if std.type(to) == 'array' then { to: to } else { to: [to] }, - // List of destinations for outgoing traffic of pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all destinations (traffic not restricted by destination). If this field is present and contains at least one item, this rule allows traffic only if the traffic matches at least one item in the to list. - withToMixin(to):: self + if std.type(to) == 'array' then { to+: to } else { to+: [to] }, - toType:: hidden.networking.v1.networkPolicyPeer, - mixin:: {}, - }, - // NetworkPolicyIngressRule describes a particular set of traffic that is allowed to the pods matched by a NetworkPolicySpec's podSelector. The traffic must match both ports and from. - networkPolicyIngressRule:: { - new():: {}, - // List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least on item, this rule allows traffic only if the traffic matches at least one item in the from list. - withFrom(from):: self + if std.type(from) == 'array' then { from: from } else { from: [from] }, - // List of sources which should be able to access the pods selected for this rule. Items in this list are combined using a logical OR operation. If this field is empty or missing, this rule matches all sources (traffic not restricted by source). If this field is present and contains at least on item, this rule allows traffic only if the traffic matches at least one item in the from list. - withFromMixin(from):: self + if std.type(from) == 'array' then { from+: from } else { from+: [from] }, - fromType:: hidden.networking.v1.networkPolicyPeer, - // List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPorts(ports):: self + if std.type(ports) == 'array' then { ports: ports } else { ports: [ports] }, - // List of ports which should be made accessible on the pods selected for this rule. Each item in this list is combined using a logical OR. If this field is empty or missing, this rule matches all ports (traffic not restricted by port). If this field is present and contains at least one item, then this rule allows traffic only if the traffic matches at least one port in the list. - withPortsMixin(ports):: self + if std.type(ports) == 'array' then { ports+: ports } else { ports+: [ports] }, - portsType:: hidden.networking.v1.networkPolicyPort, - mixin:: {}, - }, - // NetworkPolicyList is a list of NetworkPolicy objects. - networkPolicyList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.networking.v1.networkPolicy, - mixin:: {}, - }, - // NetworkPolicyPeer describes a peer to allow traffic from. Only certain combinations of fields are allowed - networkPolicyPeer:: { - new():: {}, - mixin:: { - // IPBlock defines policy on a particular IPBlock. If this field is set then neither of the other fields can be. - ipBlock:: { - local __ipBlockMixin(ipBlock) = { ipBlock+: ipBlock }, - mixinInstance(ipBlock):: __ipBlockMixin(ipBlock), - // CIDR is a string representing the IP Block Valid examples are "192.168.1.1/24" - withCidr(cidr):: self + __ipBlockMixin({ cidr: cidr }), - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExcept(except):: self + if std.type(except) == 'array' then __ipBlockMixin({ except: except }) else __ipBlockMixin({ except: [except] }), - // Except is a slice of CIDRs that should not be included within an IP Block Valid examples are "192.168.1.1/24" Except values will be rejected if they are outside the CIDR range - withExceptMixin(except):: self + if std.type(except) == 'array' then __ipBlockMixin({ except+: except }) else __ipBlockMixin({ except+: [except] }), - }, - ipBlockType:: hidden.networking.v1.ipBlock, - // Selects Namespaces using cluster-scoped labels. This field follows standard label selector semantics; if present but empty, it selects all namespaces. - // - // If PodSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects all Pods in the Namespaces selected by NamespaceSelector. - namespaceSelector:: { - local __namespaceSelectorMixin(namespaceSelector) = { namespaceSelector+: namespaceSelector }, - mixinInstance(namespaceSelector):: __namespaceSelectorMixin(namespaceSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __namespaceSelectorMixin({ matchExpressions+: matchExpressions }) else __namespaceSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __namespaceSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __namespaceSelectorMixin({ matchLabels+: matchLabels }), - }, - namespaceSelectorType:: hidden.meta.v1.labelSelector, - // This is a label selector which selects Pods. This field follows standard label selector semantics; if present but empty, it selects all pods. - // - // If NamespaceSelector is also set, then the NetworkPolicyPeer as a whole selects the Pods matching PodSelector in the Namespaces selected by NamespaceSelector. Otherwise it selects the Pods matching PodSelector in the policy's own Namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = { podSelector+: podSelector }, - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - // NetworkPolicyPort describes a port to allow traffic on - networkPolicyPort:: { - new():: {}, - // The port on the given protocol. This can either be a numerical or named port on a pod. If this field is not provided, this matches all port names and numbers. - withPort(port):: self + { port: port }, - // The protocol (TCP, UDP, or SCTP) which traffic must match. If not specified, this field defaults to TCP. - withProtocol(protocol):: self + { protocol: protocol }, - mixin:: {}, - }, - // NetworkPolicySpec provides the specification of a NetworkPolicy - networkPolicySpec:: { - new():: {}, - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgress(egress):: self + if std.type(egress) == 'array' then { egress: egress } else { egress: [egress] }, - // List of egress rules to be applied to the selected pods. Outgoing traffic is allowed if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic matches at least one egress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy limits all outgoing traffic (and serves solely to ensure that the pods it selects are isolated by default). This field is beta-level in 1.8 - withEgressMixin(egress):: self + if std.type(egress) == 'array' then { egress+: egress } else { egress+: [egress] }, - egressType:: hidden.networking.v1.networkPolicyEgressRule, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default) - withIngress(ingress):: self + if std.type(ingress) == 'array' then { ingress: ingress } else { ingress: [ingress] }, - // List of ingress rules to be applied to the selected pods. Traffic is allowed to a pod if there are no NetworkPolicies selecting the pod (and cluster policy otherwise allows the traffic), OR if the traffic source is the pod's local node, OR if the traffic matches at least one ingress rule across all of the NetworkPolicy objects whose podSelector matches the pod. If this field is empty then this NetworkPolicy does not allow any traffic (and serves solely to ensure that the pods it selects are isolated by default) - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then { ingress+: ingress } else { ingress+: [ingress] }, - ingressType:: hidden.networking.v1.networkPolicyIngressRule, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypes(policyTypes):: self + if std.type(policyTypes) == 'array' then { policyTypes: policyTypes } else { policyTypes: [policyTypes] }, - // List of rule types that the NetworkPolicy relates to. Valid options are "Ingress", "Egress", or "Ingress,Egress". If this field is not specified, it will default based on the existence of Ingress or Egress rules; policies that contain an Egress section are assumed to affect Egress, and all policies (whether or not they contain an Ingress section) are assumed to affect Ingress. If you want to write an egress-only policy, you must explicitly specify policyTypes [ "Egress" ]. Likewise, if you want to write a policy that specifies that no egress is allowed, you must specify a policyTypes value that include "Egress" (since such a policy would not include an Egress section and would otherwise default to just [ "Ingress" ]). This field is beta-level in 1.8 - withPolicyTypesMixin(policyTypes):: self + if std.type(policyTypes) == 'array' then { policyTypes+: policyTypes } else { policyTypes+: [policyTypes] }, - mixin:: { - // Selects the pods to which this NetworkPolicy object applies. The array of ingress rules is applied to any pods selected by this field. Multiple network policies can select the same set of pods. In this case, the ingress rules for each are combined additively. This field is NOT optional and follows standard label selector semantics. An empty podSelector matches all pods in this namespace. - podSelector:: { - local __podSelectorMixin(podSelector) = { podSelector+: podSelector }, - mixinInstance(podSelector):: __podSelectorMixin(podSelector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions: matchExpressions }) else __podSelectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __podSelectorMixin({ matchExpressions+: matchExpressions }) else __podSelectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __podSelectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __podSelectorMixin({ matchLabels+: matchLabels }), - }, - podSelectorType:: hidden.meta.v1.labelSelector, - }, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'networking/v1beta1' }, - // HTTPIngressPath associates a path regex with a backend. Incoming urls matching the path are forwarded to the backend. - httpIngressPath:: { - new():: {}, - // Path is an extended POSIX regex as defined by IEEE Std 1003.1, (i.e this follows the egrep/unix syntax, not the perl syntax) matched against the path of an incoming request. Currently it can contain characters disallowed from the conventional "path" part of a URL as defined by RFC 3986. Paths must begin with a '/'. If unspecified, the path defaults to a catch all sending traffic to the backend. - withPath(path):: self + { path: path }, - mixin:: { - // Backend defines the referenced service endpoint to which the traffic will be forwarded to. - backend:: { - local __backendMixin(backend) = { backend+: backend }, - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.networking.v1beta1.ingressBackend, - }, - }, - // HTTPIngressRuleValue is a list of http selectors pointing to backends. In the example: http:///? -> backend where where parts of the url correspond to RFC 3986, this resource will be used to match against everything after the last '/' and before the first '?' or '#'. - httpIngressRuleValue:: { - new():: {}, - // A collection of paths that map requests to backends. - withPaths(paths):: self + if std.type(paths) == 'array' then { paths: paths } else { paths: [paths] }, - // A collection of paths that map requests to backends. - withPathsMixin(paths):: self + if std.type(paths) == 'array' then { paths+: paths } else { paths+: [paths] }, - pathsType:: hidden.networking.v1beta1.httpIngressPath, - mixin:: {}, - }, - // IngressBackend describes all endpoints for a given service and port. - ingressBackend:: { - new():: {}, - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + { serviceName: serviceName }, - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + { servicePort: servicePort }, - mixin:: {}, - }, - // IngressList is a collection of Ingress. - ingressList:: { - new():: {}, - // Items is the list of Ingress. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of Ingress. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.networking.v1beta1.ingress, - mixin:: {}, - }, - // IngressRule represents the rules mapping the paths under a specified host to the related backend services. Incoming requests are first evaluated for a host match, then routed to the backend associated with the matching IngressRuleValue. - ingressRule:: { - new():: {}, - // Host is the fully qualified domain name of a network host, as defined by RFC 3986. Note the following deviations from the "host" part of the URI as defined in the RFC: 1. IPs are not allowed. Currently an IngressRuleValue can only apply to the - // IP in the Spec of the parent Ingress. - // 2. The `:` delimiter is not respected because ports are not allowed. - // Currently the port of an Ingress is implicitly :80 for http and - // :443 for https. - // Both these may change in the future. Incoming requests are matched against the host before the IngressRuleValue. If the host is unspecified, the Ingress routes all traffic based on the specified IngressRuleValue. - withHost(host):: self + { host: host }, - mixin:: { - http:: { - local __httpMixin(http) = { http+: http }, - mixinInstance(http):: __httpMixin(http), - // A collection of paths that map requests to backends. - withPaths(paths):: self + if std.type(paths) == 'array' then __httpMixin({ paths: paths }) else __httpMixin({ paths: [paths] }), - // A collection of paths that map requests to backends. - withPathsMixin(paths):: self + if std.type(paths) == 'array' then __httpMixin({ paths+: paths }) else __httpMixin({ paths+: [paths] }), - pathsType:: hidden.networking.v1beta1.httpIngressPath, - }, - httpType:: hidden.networking.v1beta1.httpIngressRuleValue, - }, - }, - // IngressSpec describes the Ingress the user wishes to exist. - ingressSpec:: { - new():: {}, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRules(rules):: self + if std.type(rules) == 'array' then { rules: rules } else { rules: [rules] }, - // A list of host rules used to configure the Ingress. If unspecified, or no rule matches, all traffic is sent to the default backend. - withRulesMixin(rules):: self + if std.type(rules) == 'array' then { rules+: rules } else { rules+: [rules] }, - rulesType:: hidden.networking.v1beta1.ingressRule, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTls(tls):: self + if std.type(tls) == 'array' then { tls: tls } else { tls: [tls] }, - // TLS configuration. Currently the Ingress only supports a single TLS port, 443. If multiple members of this list specify different hosts, they will be multiplexed on the same port according to the hostname specified through the SNI TLS extension, if the ingress controller fulfilling the ingress supports SNI. - withTlsMixin(tls):: self + if std.type(tls) == 'array' then { tls+: tls } else { tls+: [tls] }, - tlsType:: hidden.networking.v1beta1.ingressTls, - mixin:: { - // A default backend capable of servicing requests that don't match any rule. At least one of 'backend' or 'rules' must be specified. This field is optional to allow the loadbalancer controller or defaulting logic to specify a global default. - backend:: { - local __backendMixin(backend) = { backend+: backend }, - mixinInstance(backend):: __backendMixin(backend), - // Specifies the name of the referenced service. - withServiceName(serviceName):: self + __backendMixin({ serviceName: serviceName }), - // Specifies the port of the referenced service. - withServicePort(servicePort):: self + __backendMixin({ servicePort: servicePort }), - }, - backendType:: hidden.networking.v1beta1.ingressBackend, - }, - }, - // IngressStatus describe the current state of the Ingress. - ingressStatus:: { - new():: {}, - mixin:: { - // LoadBalancer contains the current status of the load-balancer. - loadBalancer:: { - local __loadBalancerMixin(loadBalancer) = { loadBalancer+: loadBalancer }, - mixinInstance(loadBalancer):: __loadBalancerMixin(loadBalancer), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngress(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress: ingress }) else __loadBalancerMixin({ ingress: [ingress] }), - // Ingress is a list containing ingress points for the load-balancer. Traffic intended for the service should be sent to these ingress points. - withIngressMixin(ingress):: self + if std.type(ingress) == 'array' then __loadBalancerMixin({ ingress+: ingress }) else __loadBalancerMixin({ ingress+: [ingress] }), - ingressType:: hidden.core.v1.loadBalancerIngress, - }, - loadBalancerType:: hidden.core.v1.loadBalancerStatus, - }, - }, - // IngressTLS describes the transport layer security associated with an Ingress. - ingressTls:: { - new():: {}, - // Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified. - withHosts(hosts):: self + if std.type(hosts) == 'array' then { hosts: hosts } else { hosts: [hosts] }, - // Hosts are a list of hosts included in the TLS certificate. The values in this list must match the name/s used in the tlsSecret. Defaults to the wildcard host setting for the loadbalancer controller fulfilling this Ingress, if left unspecified. - withHostsMixin(hosts):: self + if std.type(hosts) == 'array' then { hosts+: hosts } else { hosts+: [hosts] }, - // SecretName is the name of the secret used to terminate SSL traffic on 443. Field is left optional to allow SSL routing based on SNI hostname alone. If the SNI host in a listener conflicts with the "Host" header field used by an IngressRule, the SNI host is used for termination and value of the Host header is used for routing. - withSecretName(secretName):: self + { secretName: secretName }, - mixin:: {}, - }, - }, - }, - node:: { - v1beta1:: { - local apiVersion = { apiVersion: 'node/v1beta1' }, - // RuntimeClassList is a list of RuntimeClass objects. - runtimeClassList:: { - new():: {}, - // Items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.node.v1beta1.runtimeClass, - mixin:: {}, - }, - }, - }, - policy:: { - v1beta1:: { - local apiVersion = { apiVersion: 'policy/v1beta1' }, - // AllowedCSIDriver represents a single inline CSI Driver that is allowed to be used. - allowedCsiDriver:: { - new():: {}, - // Name is the registered name of the CSI driver - withName(name):: self + { name: name }, - mixin:: {}, - }, - // AllowedFlexVolume represents a single Flexvolume that is allowed to be used. - allowedFlexVolume:: { - new():: {}, - // driver is the name of the Flexvolume driver. - withDriver(driver):: self + { driver: driver }, - mixin:: {}, - }, - // AllowedHostPath defines the host volume conditions that will be enabled by a policy for pods to use. It requires the path prefix to be defined. - allowedHostPath:: { - new():: {}, - // pathPrefix is the path prefix that the host volume must match. It does not support `*`. Trailing slashes are trimmed when validating the path prefix with a host path. - // - // Examples: `/foo` would allow `/foo`, `/foo/` and `/foo/bar` `/foo` would not allow `/food` or `/etc/foo` - withPathPrefix(pathPrefix):: self + { pathPrefix: pathPrefix }, - // when set to true, will allow host volumes matching the pathPrefix only if all volume mounts are readOnly. - withReadOnly(readOnly):: self + { readOnly: readOnly }, - mixin:: {}, - }, - // FSGroupStrategyOptions defines the strategy type and options used to create the strategy. - fsGroupStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // HostPortRange defines a range of host ports that will be enabled by a policy for pods to use. It requires both the start and end to be defined. - hostPortRange:: { - new():: {}, - // max is the end of the range, inclusive. - withMax(max):: self + { max: max }, - // min is the start of the range, inclusive. - withMin(min):: self + { min: min }, - mixin:: {}, - }, - // IDRange provides a min/max of an allowed range of IDs. - idRange:: { - new():: {}, - // max is the end of the range, inclusive. - withMax(max):: self + { max: max }, - // min is the start of the range, inclusive. - withMin(min):: self + { min: min }, - mixin:: {}, - }, - // PodDisruptionBudgetList is a collection of PodDisruptionBudgets. - podDisruptionBudgetList:: { - new():: {}, - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.policy.v1beta1.podDisruptionBudget, - mixin:: {}, - }, - // PodDisruptionBudgetSpec is a description of a PodDisruptionBudget. - podDisruptionBudgetSpec:: { - new():: {}, - // An eviction is allowed if at most "maxUnavailable" pods selected by "selector" are unavailable after the eviction, i.e. even in absence of the evicted pod. For example, one can prevent all voluntary evictions by specifying 0. This is a mutually exclusive setting with "minAvailable". - withMaxUnavailable(maxUnavailable):: self + { maxUnavailable: maxUnavailable }, - // An eviction is allowed if at least "minAvailable" pods selected by "selector" will still be available after the eviction, i.e. even in the absence of the evicted pod. So for example you can prevent all voluntary evictions by specifying "100%". - withMinAvailable(minAvailable):: self + { minAvailable: minAvailable }, - mixin:: { - // Label query over pods whose evictions are managed by the disruption budget. - selector:: { - local __selectorMixin(selector) = { selector+: selector }, - mixinInstance(selector):: __selectorMixin(selector), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressions(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions: matchExpressions }) else __selectorMixin({ matchExpressions: [matchExpressions] }), - // matchExpressions is a list of label selector requirements. The requirements are ANDed. - withMatchExpressionsMixin(matchExpressions):: self + if std.type(matchExpressions) == 'array' then __selectorMixin({ matchExpressions+: matchExpressions }) else __selectorMixin({ matchExpressions+: [matchExpressions] }), - matchExpressionsType:: hidden.meta.v1.labelSelectorRequirement, - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabels(matchLabels):: self + __selectorMixin({ matchLabels: matchLabels }), - // matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is "key", the operator is "In", and the values array contains only "value". The requirements are ANDed. - withMatchLabelsMixin(matchLabels):: self + __selectorMixin({ matchLabels+: matchLabels }), - }, - selectorType:: hidden.meta.v1.labelSelector, - }, - }, - // PodDisruptionBudgetStatus represents information about the status of a PodDisruptionBudget. Status may trail the actual state of a system. - podDisruptionBudgetStatus:: { - new():: {}, - // current number of healthy pods - withCurrentHealthy(currentHealthy):: self + { currentHealthy: currentHealthy }, - // minimum desired number of healthy pods - withDesiredHealthy(desiredHealthy):: self + { desiredHealthy: desiredHealthy }, - // DisruptedPods contains information about pods whose eviction was processed by the API server eviction subresource handler but has not yet been observed by the PodDisruptionBudget controller. A pod will be in this map from the time when the API server processed the eviction request to the time when the pod is seen by PDB controller as having been marked for deletion (or after a timeout). The key in the map is the name of the pod and the value is the time when the API server processed the eviction request. If the deletion didn't occur and a pod is still there it will be removed from the list automatically by PodDisruptionBudget controller after some time. If everything goes smooth this map should be empty for the most of the time. Large number of entries in the map may indicate problems with pod deletions. - withDisruptedPods(disruptedPods):: self + { disruptedPods: disruptedPods }, - // DisruptedPods contains information about pods whose eviction was processed by the API server eviction subresource handler but has not yet been observed by the PodDisruptionBudget controller. A pod will be in this map from the time when the API server processed the eviction request to the time when the pod is seen by PDB controller as having been marked for deletion (or after a timeout). The key in the map is the name of the pod and the value is the time when the API server processed the eviction request. If the deletion didn't occur and a pod is still there it will be removed from the list automatically by PodDisruptionBudget controller after some time. If everything goes smooth this map should be empty for the most of the time. Large number of entries in the map may indicate problems with pod deletions. - withDisruptedPodsMixin(disruptedPods):: self + { disruptedPods+: disruptedPods }, - // Number of pod disruptions that are currently allowed. - withDisruptionsAllowed(disruptionsAllowed):: self + { disruptionsAllowed: disruptionsAllowed }, - // total number of pods counted by this disruption budget - withExpectedPods(expectedPods):: self + { expectedPods: expectedPods }, - // Most recent generation observed when updating this PDB status. PodDisruptionsAllowed and other status informatio is valid only if observedGeneration equals to PDB's object generation. - withObservedGeneration(observedGeneration):: self + { observedGeneration: observedGeneration }, - mixin:: {}, - }, - // PodSecurityPolicyList is a list of PodSecurityPolicy objects. - podSecurityPolicyList:: { - new():: {}, - // items is a list of schema objects. - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is a list of schema objects. - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.policy.v1beta1.podSecurityPolicy, - mixin:: {}, - }, - // PodSecurityPolicySpec defines the policy enforced. - podSecurityPolicySpec:: { - new():: {}, - // allowPrivilegeEscalation determines if a pod can request to allow privilege escalation. If unspecified, defaults to true. - withAllowPrivilegeEscalation(allowPrivilegeEscalation):: self + { allowPrivilegeEscalation: allowPrivilegeEscalation }, - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDrivers(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then { allowedCSIDrivers: allowedCsiDrivers } else { allowedCSIDrivers: [allowedCsiDrivers] }, - // AllowedCSIDrivers is a whitelist of inline CSI drivers that must be explicitly set to be embedded within a pod spec. An empty value means no CSI drivers can run inline within a pod spec. - withAllowedCsiDriversMixin(allowedCsiDrivers):: self + if std.type(allowedCsiDrivers) == 'array' then { allowedCSIDrivers+: allowedCsiDrivers } else { allowedCSIDrivers+: [allowedCsiDrivers] }, - allowedCSIDriversType:: hidden.policy.v1beta1.allowedCsiDriver, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilities(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then { allowedCapabilities: allowedCapabilities } else { allowedCapabilities: [allowedCapabilities] }, - // allowedCapabilities is a list of capabilities that can be requested to add to the container. Capabilities in this field may be added at the pod author's discretion. You must not list a capability in both allowedCapabilities and requiredDropCapabilities. - withAllowedCapabilitiesMixin(allowedCapabilities):: self + if std.type(allowedCapabilities) == 'array' then { allowedCapabilities+: allowedCapabilities } else { allowedCapabilities+: [allowedCapabilities] }, - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumes(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then { allowedFlexVolumes: allowedFlexVolumes } else { allowedFlexVolumes: [allowedFlexVolumes] }, - // allowedFlexVolumes is a whitelist of allowed Flexvolumes. Empty or nil indicates that all Flexvolumes may be used. This parameter is effective only when the usage of the Flexvolumes is allowed in the "volumes" field. - withAllowedFlexVolumesMixin(allowedFlexVolumes):: self + if std.type(allowedFlexVolumes) == 'array' then { allowedFlexVolumes+: allowedFlexVolumes } else { allowedFlexVolumes+: [allowedFlexVolumes] }, - allowedFlexVolumesType:: hidden.policy.v1beta1.allowedFlexVolume, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPaths(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then { allowedHostPaths: allowedHostPaths } else { allowedHostPaths: [allowedHostPaths] }, - // allowedHostPaths is a white list of allowed host paths. Empty indicates that all host paths may be used. - withAllowedHostPathsMixin(allowedHostPaths):: self + if std.type(allowedHostPaths) == 'array' then { allowedHostPaths+: allowedHostPaths } else { allowedHostPaths+: [allowedHostPaths] }, - allowedHostPathsType:: hidden.policy.v1beta1.allowedHostPath, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypes(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then { allowedProcMountTypes: allowedProcMountTypes } else { allowedProcMountTypes: [allowedProcMountTypes] }, - // AllowedProcMountTypes is a whitelist of allowed ProcMountTypes. Empty or nil indicates that only the DefaultProcMountType may be used. This requires the ProcMountType feature flag to be enabled. - withAllowedProcMountTypesMixin(allowedProcMountTypes):: self + if std.type(allowedProcMountTypes) == 'array' then { allowedProcMountTypes+: allowedProcMountTypes } else { allowedProcMountTypes+: [allowedProcMountTypes] }, - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctls(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then { allowedUnsafeSysctls: allowedUnsafeSysctls } else { allowedUnsafeSysctls: [allowedUnsafeSysctls] }, - // allowedUnsafeSysctls is a list of explicitly allowed unsafe sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of allowed sysctls. Single * means all unsafe sysctls are allowed. Kubelet has to whitelist all allowed unsafe sysctls explicitly to avoid rejection. - // - // Examples: e.g. "foo/*" allows "foo/bar", "foo/baz", etc. e.g. "foo.*" allows "foo.bar", "foo.baz", etc. - withAllowedUnsafeSysctlsMixin(allowedUnsafeSysctls):: self + if std.type(allowedUnsafeSysctls) == 'array' then { allowedUnsafeSysctls+: allowedUnsafeSysctls } else { allowedUnsafeSysctls+: [allowedUnsafeSysctls] }, - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilities(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then { defaultAddCapabilities: defaultAddCapabilities } else { defaultAddCapabilities: [defaultAddCapabilities] }, - // defaultAddCapabilities is the default set of capabilities that will be added to the container unless the pod spec specifically drops the capability. You may not list a capability in both defaultAddCapabilities and requiredDropCapabilities. Capabilities added here are implicitly allowed, and need not be included in the allowedCapabilities list. - withDefaultAddCapabilitiesMixin(defaultAddCapabilities):: self + if std.type(defaultAddCapabilities) == 'array' then { defaultAddCapabilities+: defaultAddCapabilities } else { defaultAddCapabilities+: [defaultAddCapabilities] }, - // defaultAllowPrivilegeEscalation controls the default setting for whether a process can gain more privileges than its parent process. - withDefaultAllowPrivilegeEscalation(defaultAllowPrivilegeEscalation):: self + { defaultAllowPrivilegeEscalation: defaultAllowPrivilegeEscalation }, - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctls(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then { forbiddenSysctls: forbiddenSysctls } else { forbiddenSysctls: [forbiddenSysctls] }, - // forbiddenSysctls is a list of explicitly forbidden sysctls, defaults to none. Each entry is either a plain sysctl name or ends in "*" in which case it is considered as a prefix of forbidden sysctls. Single * means all sysctls are forbidden. - // - // Examples: e.g. "foo/*" forbids "foo/bar", "foo/baz", etc. e.g. "foo.*" forbids "foo.bar", "foo.baz", etc. - withForbiddenSysctlsMixin(forbiddenSysctls):: self + if std.type(forbiddenSysctls) == 'array' then { forbiddenSysctls+: forbiddenSysctls } else { forbiddenSysctls+: [forbiddenSysctls] }, - // hostIPC determines if the policy allows the use of HostIPC in the pod spec. - withHostIpc(hostIpc):: self + { hostIPC: hostIpc }, - // hostNetwork determines if the policy allows the use of HostNetwork in the pod spec. - withHostNetwork(hostNetwork):: self + { hostNetwork: hostNetwork }, - // hostPID determines if the policy allows the use of HostPID in the pod spec. - withHostPid(hostPid):: self + { hostPID: hostPid }, - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPorts(hostPorts):: self + if std.type(hostPorts) == 'array' then { hostPorts: hostPorts } else { hostPorts: [hostPorts] }, - // hostPorts determines which host port ranges are allowed to be exposed. - withHostPortsMixin(hostPorts):: self + if std.type(hostPorts) == 'array' then { hostPorts+: hostPorts } else { hostPorts+: [hostPorts] }, - hostPortsType:: hidden.policy.v1beta1.hostPortRange, - // privileged determines if a pod can request to be run as privileged. - withPrivileged(privileged):: self + { privileged: privileged }, - // readOnlyRootFilesystem when set to true will force containers to run with a read only root file system. If the container specifically requests to run with a non-read only root file system the PSP should deny the pod. If set to false the container may run with a read only root file system if it wishes but it will not be forced to. - withReadOnlyRootFilesystem(readOnlyRootFilesystem):: self + { readOnlyRootFilesystem: readOnlyRootFilesystem }, - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilities(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then { requiredDropCapabilities: requiredDropCapabilities } else { requiredDropCapabilities: [requiredDropCapabilities] }, - // requiredDropCapabilities are the capabilities that will be dropped from the container. These are required to be dropped and cannot be added. - withRequiredDropCapabilitiesMixin(requiredDropCapabilities):: self + if std.type(requiredDropCapabilities) == 'array' then { requiredDropCapabilities+: requiredDropCapabilities } else { requiredDropCapabilities+: [requiredDropCapabilities] }, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumes(volumes):: self + if std.type(volumes) == 'array' then { volumes: volumes } else { volumes: [volumes] }, - // volumes is a white list of allowed volume plugins. Empty indicates that no volumes may be used. To allow all volumes you may use '*'. - withVolumesMixin(volumes):: self + if std.type(volumes) == 'array' then { volumes+: volumes } else { volumes+: [volumes] }, - mixin:: { - // fsGroup is the strategy that will dictate what fs group is used by the SecurityContext. - fsGroup:: { - local __fsGroupMixin(fsGroup) = { fsGroup+: fsGroup }, - mixinInstance(fsGroup):: __fsGroupMixin(fsGroup), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges: ranges }) else __fsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of fs groups. If you would like to force a single fs group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __fsGroupMixin({ ranges+: ranges }) else __fsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what FSGroup is used in the SecurityContext. - withRule(rule):: self + __fsGroupMixin({ rule: rule }), - }, - fsGroupType:: hidden.policy.v1beta1.fsGroupStrategyOptions, - // RunAsGroup is the strategy that will dictate the allowable RunAsGroup values that may be set. If this field is omitted, the pod's RunAsGroup can take any value. This field requires the RunAsGroup feature gate to be enabled. - runAsGroup:: { - local __runAsGroupMixin(runAsGroup) = { runAsGroup+: runAsGroup }, - mixinInstance(runAsGroup):: __runAsGroupMixin(runAsGroup), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges: ranges }) else __runAsGroupMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsGroupMixin({ ranges+: ranges }) else __runAsGroupMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + __runAsGroupMixin({ rule: rule }), - }, - runAsGroupType:: hidden.policy.v1beta1.runAsGroupStrategyOptions, - // runAsUser is the strategy that will dictate the allowable RunAsUser values that may be set. - runAsUser:: { - local __runAsUserMixin(runAsUser) = { runAsUser+: runAsUser }, - mixinInstance(runAsUser):: __runAsUserMixin(runAsUser), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges: ranges }) else __runAsUserMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __runAsUserMixin({ ranges+: ranges }) else __runAsUserMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + __runAsUserMixin({ rule: rule }), - }, - runAsUserType:: hidden.policy.v1beta1.runAsUserStrategyOptions, - // seLinux is the strategy that will dictate the allowable labels that may be set. - seLinux:: { - local __seLinuxMixin(seLinux) = { seLinux+: seLinux }, - mixinInstance(seLinux):: __seLinuxMixin(seLinux), - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + __seLinuxMixin({ rule: rule }), - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = __seLinuxMixin({ seLinuxOptions+: seLinuxOptions }), - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - seLinuxType:: hidden.policy.v1beta1.seLinuxStrategyOptions, - // supplementalGroups is the strategy that will dictate what supplemental groups are used by the SecurityContext. - supplementalGroups:: { - local __supplementalGroupsMixin(supplementalGroups) = { supplementalGroups+: supplementalGroups }, - mixinInstance(supplementalGroups):: __supplementalGroupsMixin(supplementalGroups), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges: ranges }) else __supplementalGroupsMixin({ ranges: [ranges] }), - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then __supplementalGroupsMixin({ ranges+: ranges }) else __supplementalGroupsMixin({ ranges+: [ranges] }), - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + __supplementalGroupsMixin({ rule: rule }), - }, - supplementalGroupsType:: hidden.policy.v1beta1.supplementalGroupsStrategyOptions, - }, - }, - // RunAsGroupStrategyOptions defines the strategy type and any options used to create the strategy. - runAsGroupStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of gids that may be used. If you would like to force a single gid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsGroup values that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // RunAsUserStrategyOptions defines the strategy type and any options used to create the strategy. - runAsUserStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of uids that may be used. If you would like to force a single uid then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate the allowable RunAsUser values that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - // SELinuxStrategyOptions defines the strategy type and any options used to create the strategy. - seLinuxStrategyOptions:: { - new():: {}, - // rule is the strategy that will dictate the allowable labels that may be set. - withRule(rule):: self + { rule: rule }, - mixin:: { - // seLinuxOptions required to run as; required for MustRunAs More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ - seLinuxOptions:: { - local __seLinuxOptionsMixin(seLinuxOptions) = { seLinuxOptions+: seLinuxOptions }, - mixinInstance(seLinuxOptions):: __seLinuxOptionsMixin(seLinuxOptions), - // Level is SELinux level label that applies to the container. - withLevel(level):: self + __seLinuxOptionsMixin({ level: level }), - // Role is a SELinux role label that applies to the container. - withRole(role):: self + __seLinuxOptionsMixin({ role: role }), - // Type is a SELinux type label that applies to the container. - withType(type):: self + __seLinuxOptionsMixin({ type: type }), - // User is a SELinux user label that applies to the container. - withUser(user):: self + __seLinuxOptionsMixin({ user: user }), - }, - seLinuxOptionsType:: hidden.core.v1.seLinuxOptions, - }, - }, - // SupplementalGroupsStrategyOptions defines the strategy type and options used to create the strategy. - supplementalGroupsStrategyOptions:: { - new():: {}, - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRanges(ranges):: self + if std.type(ranges) == 'array' then { ranges: ranges } else { ranges: [ranges] }, - // ranges are the allowed ranges of supplemental groups. If you would like to force a single supplemental group then supply a single range with the same start and end. Required for MustRunAs. - withRangesMixin(ranges):: self + if std.type(ranges) == 'array' then { ranges+: ranges } else { ranges+: [ranges] }, - rangesType:: hidden.policy.v1beta1.idRange, - // rule is the strategy that will dictate what supplemental groups is used in the SecurityContext. - withRule(rule):: self + { rule: rule }, - mixin:: {}, - }, - }, - }, - rbac:: { - v1:: { - local apiVersion = { apiVersion: 'rbac/v1' }, - // AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole - aggregationRule:: { - new():: {}, - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectors(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then { clusterRoleSelectors: clusterRoleSelectors } else { clusterRoleSelectors: [clusterRoleSelectors] }, - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectorsMixin(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then { clusterRoleSelectors+: clusterRoleSelectors } else { clusterRoleSelectors+: [clusterRoleSelectors] }, - clusterRoleSelectorsType:: hidden.meta.v1.labelSelector, - mixin:: {}, - }, - // ClusterRoleBindingList is a collection of ClusterRoleBindings - clusterRoleBindingList:: { - new():: {}, - // Items is a list of ClusterRoleBindings - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of ClusterRoleBindings - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1.clusterRoleBinding, - mixin:: {}, - }, - // ClusterRoleList is a collection of ClusterRoles - clusterRoleList:: { - new():: {}, - // Items is a list of ClusterRoles - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of ClusterRoles - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1.clusterRole, - mixin:: {}, - }, - // PolicyRule holds information that describes a policy rule, but does not contain information about who the rule applies to or which namespace the rule applies to. - policyRule:: { - new():: {}, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. - withApiGroups(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups: apiGroups } else { apiGroups: [apiGroups] }, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. - withApiGroupsMixin(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups+: apiGroups } else { apiGroups+: [apiGroups] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. - withNonResourceUrls(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs: nonResourceUrls } else { nonResourceURLs: [nonResourceUrls] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. - withNonResourceUrlsMixin(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs+: nonResourceUrls } else { nonResourceURLs+: [nonResourceUrls] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. - withResourceNames(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames: resourceNames } else { resourceNames: [resourceNames] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. - withResourceNamesMixin(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames+: resourceNames } else { resourceNames+: [resourceNames] }, - // Resources is a list of resources this rule applies to. ResourceAll represents all resources. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // Resources is a list of resources this rule applies to. ResourceAll represents all resources. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - // Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // RoleBindingList is a collection of RoleBindings - roleBindingList:: { - new():: {}, - // Items is a list of RoleBindings - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of RoleBindings - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1.roleBinding, - mixin:: {}, - }, - // RoleList is a collection of Roles - roleList:: { - new():: {}, - // Items is a list of Roles - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of Roles - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1.role, - mixin:: {}, - }, - // RoleRef contains information that points to the role being used - roleRef:: { - new():: {}, - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + { apiGroup: apiGroup }, - // Kind is the type of resource being referenced - withKind(kind):: self + { kind: kind }, - // Name is the name of resource being referenced - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Subject contains a reference to the object or user identities a role binding applies to. This can either hold a direct API object reference, or a value for non-objects such as user and group names. - subject:: { - new():: {}, - // APIGroup holds the API group of the referenced subject. Defaults to "" for ServiceAccount subjects. Defaults to "rbac.authorization.k8s.io" for User and Group subjects. - withApiGroup(apiGroup):: self + { apiGroup: apiGroup }, - // Kind of object being referenced. Values defined by this API group are "User", "Group", and "ServiceAccount". If the Authorizer does not recognized the kind value, the Authorizer should report an error. - withKind(kind):: self + { kind: kind }, - // Name of the object being referenced. - withName(name):: self + { name: name }, - // Namespace of the referenced object. If the object kind is non-namespace, such as "User" or "Group", and this value is not empty the Authorizer should report an error. - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'rbac/v1beta1' }, - // AggregationRule describes how to locate ClusterRoles to aggregate into the ClusterRole - aggregationRule:: { - new():: {}, - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectors(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then { clusterRoleSelectors: clusterRoleSelectors } else { clusterRoleSelectors: [clusterRoleSelectors] }, - // ClusterRoleSelectors holds a list of selectors which will be used to find ClusterRoles and create the rules. If any of the selectors match, then the ClusterRole's permissions will be added - withClusterRoleSelectorsMixin(clusterRoleSelectors):: self + if std.type(clusterRoleSelectors) == 'array' then { clusterRoleSelectors+: clusterRoleSelectors } else { clusterRoleSelectors+: [clusterRoleSelectors] }, - clusterRoleSelectorsType:: hidden.meta.v1.labelSelector, - mixin:: {}, - }, - // ClusterRoleBindingList is a collection of ClusterRoleBindings - clusterRoleBindingList:: { - new():: {}, - // Items is a list of ClusterRoleBindings - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of ClusterRoleBindings - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1beta1.clusterRoleBinding, - mixin:: {}, - }, - // ClusterRoleList is a collection of ClusterRoles - clusterRoleList:: { - new():: {}, - // Items is a list of ClusterRoles - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of ClusterRoles - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1beta1.clusterRole, - mixin:: {}, - }, - // PolicyRule holds information that describes a policy rule, but does not contain information about who the rule applies to or which namespace the rule applies to. - policyRule:: { - new():: {}, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. - withApiGroups(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups: apiGroups } else { apiGroups: [apiGroups] }, - // APIGroups is the name of the APIGroup that contains the resources. If multiple API groups are specified, any action requested against one of the enumerated resources in any API group will be allowed. - withApiGroupsMixin(apiGroups):: self + if std.type(apiGroups) == 'array' then { apiGroups+: apiGroups } else { apiGroups+: [apiGroups] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. - withNonResourceUrls(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs: nonResourceUrls } else { nonResourceURLs: [nonResourceUrls] }, - // NonResourceURLs is a set of partial urls that a user should have access to. *s are allowed, but only as the full, final step in the path Since non-resource URLs are not namespaced, this field is only applicable for ClusterRoles referenced from a ClusterRoleBinding. Rules can either apply to API resources (such as "pods" or "secrets") or non-resource URL paths (such as "/api"), but not both. - withNonResourceUrlsMixin(nonResourceUrls):: self + if std.type(nonResourceUrls) == 'array' then { nonResourceURLs+: nonResourceUrls } else { nonResourceURLs+: [nonResourceUrls] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. - withResourceNames(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames: resourceNames } else { resourceNames: [resourceNames] }, - // ResourceNames is an optional white list of names that the rule applies to. An empty set means that everything is allowed. - withResourceNamesMixin(resourceNames):: self + if std.type(resourceNames) == 'array' then { resourceNames+: resourceNames } else { resourceNames+: [resourceNames] }, - // Resources is a list of resources this rule applies to. '*' represents all resources in the specified apiGroups. '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups. - withResources(resources):: self + if std.type(resources) == 'array' then { resources: resources } else { resources: [resources] }, - // Resources is a list of resources this rule applies to. '*' represents all resources in the specified apiGroups. '*/foo' represents the subresource 'foo' for all resources in the specified apiGroups. - withResourcesMixin(resources):: self + if std.type(resources) == 'array' then { resources+: resources } else { resources+: [resources] }, - // Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds. - withVerbs(verbs):: self + if std.type(verbs) == 'array' then { verbs: verbs } else { verbs: [verbs] }, - // Verbs is a list of Verbs that apply to ALL the ResourceKinds and AttributeRestrictions contained in this rule. VerbAll represents all kinds. - withVerbsMixin(verbs):: self + if std.type(verbs) == 'array' then { verbs+: verbs } else { verbs+: [verbs] }, - mixin:: {}, - }, - // RoleBindingList is a collection of RoleBindings - roleBindingList:: { - new():: {}, - // Items is a list of RoleBindings - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of RoleBindings - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1beta1.roleBinding, - mixin:: {}, - }, - // RoleList is a collection of Roles - roleList:: { - new():: {}, - // Items is a list of Roles - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is a list of Roles - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.rbac.v1beta1.role, - mixin:: {}, - }, - // RoleRef contains information that points to the role being used - roleRef:: { - new():: {}, - // APIGroup is the group for the resource being referenced - withApiGroup(apiGroup):: self + { apiGroup: apiGroup }, - // Kind is the type of resource being referenced - withKind(kind):: self + { kind: kind }, - // Name is the name of resource being referenced - withName(name):: self + { name: name }, - mixin:: {}, - }, - // Subject contains a reference to the object or user identities a role binding applies to. This can either hold a direct API object reference, or a value for non-objects such as user and group names. - subject:: { - new():: {}, - // APIGroup holds the API group of the referenced subject. Defaults to "" for ServiceAccount subjects. Defaults to "rbac.authorization.k8s.io" for User and Group subjects. - withApiGroup(apiGroup):: self + { apiGroup: apiGroup }, - // Kind of object being referenced. Values defined by this API group are "User", "Group", and "ServiceAccount". If the Authorizer does not recognized the kind value, the Authorizer should report an error. - withKind(kind):: self + { kind: kind }, - // Name of the object being referenced. - withName(name):: self + { name: name }, - // Namespace of the referenced object. If the object kind is non-namespace, such as "User" or "Group", and this value is not empty the Authorizer should report an error. - withNamespace(namespace):: self + { namespace: namespace }, - mixin:: {}, - }, - }, - }, - scheduling:: { - v1:: { - local apiVersion = { apiVersion: 'scheduling/v1' }, - // PriorityClassList is a collection of priority classes. - priorityClassList:: { - new():: {}, - // items is the list of PriorityClasses - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of PriorityClasses - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.scheduling.v1.priorityClass, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'scheduling/v1beta1' }, - // PriorityClassList is a collection of priority classes. - priorityClassList:: { - new():: {}, - // items is the list of PriorityClasses - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of PriorityClasses - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.scheduling.v1beta1.priorityClass, - mixin:: {}, - }, - }, - }, - storage:: { - v1:: { - local apiVersion = { apiVersion: 'storage/v1' }, - // StorageClassList is a collection of storage classes. - storageClassList:: { - new():: {}, - // Items is the list of StorageClasses - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of StorageClasses - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1.storageClass, - mixin:: {}, - }, - // VolumeAttachmentList is a collection of VolumeAttachment objects. - volumeAttachmentList:: { - new():: {}, - // Items is the list of VolumeAttachments - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of VolumeAttachments - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1.volumeAttachment, - mixin:: {}, - }, - // VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set. - volumeAttachmentSource:: { - new():: {}, - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + { persistentVolumeName: persistentVolumeName }, - mixin:: {}, - }, - // VolumeAttachmentSpec is the specification of a VolumeAttachment request. - volumeAttachmentSpec:: { - new():: {}, - // Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName(). - withAttacher(attacher):: self + { attacher: attacher }, - // The node that the volume should be attached to. - withNodeName(nodeName):: self + { nodeName: nodeName }, - mixin:: { - // Source represents the volume that should be attached. - source:: { - local __sourceMixin(source) = { source+: source }, - mixinInstance(source):: __sourceMixin(source), - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + __sourceMixin({ persistentVolumeName: persistentVolumeName }), - }, - sourceType:: hidden.storage.v1.volumeAttachmentSource, - }, - }, - // VolumeAttachmentStatus is the status of a VolumeAttachment request. - volumeAttachmentStatus:: { - new():: {}, - // Indicates the volume is successfully attached. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttached(attached):: self + { attached: attached }, - // Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttachmentMetadata(attachmentMetadata):: self + { attachmentMetadata: attachmentMetadata }, - // Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttachmentMetadataMixin(attachmentMetadata):: self + { attachmentMetadata+: attachmentMetadata }, - mixin:: { - // The last error encountered during attach operation, if any. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - attachError:: { - local __attachErrorMixin(attachError) = { attachError+: attachError }, - mixinInstance(attachError):: __attachErrorMixin(attachError), - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + __attachErrorMixin({ message: message }), - // Time the error was encountered. - withTime(time):: self + __attachErrorMixin({ time: time }), - }, - attachErrorType:: hidden.storage.v1.volumeError, - // The last error encountered during detach operation, if any. This field must only be set by the entity completing the detach operation, i.e. the external-attacher. - detachError:: { - local __detachErrorMixin(detachError) = { detachError+: detachError }, - mixinInstance(detachError):: __detachErrorMixin(detachError), - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + __detachErrorMixin({ message: message }), - // Time the error was encountered. - withTime(time):: self + __detachErrorMixin({ time: time }), - }, - detachErrorType:: hidden.storage.v1.volumeError, - }, - }, - // VolumeError captures an error encountered during a volume operation. - volumeError:: { - new():: {}, - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + { message: message }, - // Time the error was encountered. - withTime(time):: self + { time: time }, - mixin:: {}, - }, - }, - v1beta1:: { - local apiVersion = { apiVersion: 'storage/v1beta1' }, - // CSIDriverList is a collection of CSIDriver objects. - csiDriverList:: { - new():: {}, - // items is the list of CSIDriver - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of CSIDriver - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1beta1.csiDriver, - mixin:: {}, - }, - // CSIDriverSpec is the specification of a CSIDriver. - csiDriverSpec:: { - new():: {}, - // attachRequired indicates this CSI volume driver requires an attach operation (because it implements the CSI ControllerPublishVolume() method), and that the Kubernetes attach detach controller should call the attach volume interface which checks the volumeattachment status and waits until the volume is attached before proceeding to mounting. The CSI external-attacher coordinates with CSI volume driver and updates the volumeattachment status when the attach operation is complete. If the CSIDriverRegistry feature gate is enabled and the value is specified to false, the attach operation will be skipped. Otherwise the attach operation will be called. - withAttachRequired(attachRequired):: self + { attachRequired: attachRequired }, - // If set to true, podInfoOnMount indicates this CSI volume driver requires additional pod information (like podName, podUID, etc.) during mount operations. If set to false, pod information will not be passed on mount. Default is false. The CSI driver specifies podInfoOnMount as part of driver deployment. If true, Kubelet will pass pod information as VolumeContext in the CSI NodePublishVolume() calls. The CSI driver is responsible for parsing and validating the information passed in as VolumeContext. The following VolumeConext will be passed if podInfoOnMount is set to true. This list might grow, but the prefix will be used. "csi.storage.k8s.io/pod.name": pod.Name "csi.storage.k8s.io/pod.namespace": pod.Namespace "csi.storage.k8s.io/pod.uid": string(pod.UID) - withPodInfoOnMount(podInfoOnMount):: self + { podInfoOnMount: podInfoOnMount }, - mixin:: {}, - }, - // CSINodeDriver holds information about the specification of one CSI driver installed on a node - csiNodeDriver:: { - new():: {}, - // This is the name of the CSI driver that this object refers to. This MUST be the same name returned by the CSI GetPluginName() call for that driver. - withName(name):: self + { name: name }, - // nodeID of the node from the driver point of view. This field enables Kubernetes to communicate with storage systems that do not share the same nomenclature for nodes. For example, Kubernetes may refer to a given node as "node1", but the storage system may refer to the same node as "nodeA". When Kubernetes issues a command to the storage system to attach a volume to a specific node, it can use this field to refer to the node name using the ID that the storage system will understand, e.g. "nodeA" instead of "node1". This field is required. - withNodeId(nodeId):: self + { nodeID: nodeId }, - // topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology. - withTopologyKeys(topologyKeys):: self + if std.type(topologyKeys) == 'array' then { topologyKeys: topologyKeys } else { topologyKeys: [topologyKeys] }, - // topologyKeys is the list of keys supported by the driver. When a driver is initialized on a cluster, it provides a set of topology keys that it understands (e.g. "company.com/zone", "company.com/region"). When a driver is initialized on a node, it provides the same topology keys along with values. Kubelet will expose these topology keys as labels on its own node object. When Kubernetes does topology aware provisioning, it can use this list to determine which labels it should retrieve from the node object and pass back to the driver. It is possible for different nodes to use different topology keys. This can be empty if driver does not support topology. - withTopologyKeysMixin(topologyKeys):: self + if std.type(topologyKeys) == 'array' then { topologyKeys+: topologyKeys } else { topologyKeys+: [topologyKeys] }, - mixin:: {}, - }, - // CSINodeList is a collection of CSINode objects. - csiNodeList:: { - new():: {}, - // items is the list of CSINode - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // items is the list of CSINode - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1beta1.csiNode, - mixin:: {}, - }, - // CSINodeSpec holds information about the specification of all CSI drivers installed on a node - csiNodeSpec:: { - new():: {}, - // drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty. - withDrivers(drivers):: self + if std.type(drivers) == 'array' then { drivers: drivers } else { drivers: [drivers] }, - // drivers is a list of information of all CSI Drivers existing on a node. If all drivers in the list are uninstalled, this can become empty. - withDriversMixin(drivers):: self + if std.type(drivers) == 'array' then { drivers+: drivers } else { drivers+: [drivers] }, - driversType:: hidden.storage.v1beta1.csiNodeDriver, - mixin:: {}, - }, - // StorageClassList is a collection of storage classes. - storageClassList:: { - new():: {}, - // Items is the list of StorageClasses - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of StorageClasses - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1beta1.storageClass, - mixin:: {}, - }, - // VolumeAttachmentList is a collection of VolumeAttachment objects. - volumeAttachmentList:: { - new():: {}, - // Items is the list of VolumeAttachments - withItems(items):: self + if std.type(items) == 'array' then { items: items } else { items: [items] }, - // Items is the list of VolumeAttachments - withItemsMixin(items):: self + if std.type(items) == 'array' then { items+: items } else { items+: [items] }, - itemsType:: hidden.storage.v1beta1.volumeAttachment, - mixin:: {}, - }, - // VolumeAttachmentSource represents a volume that should be attached. Right now only PersistenVolumes can be attached via external attacher, in future we may allow also inline volumes in pods. Exactly one member can be set. - volumeAttachmentSource:: { - new():: {}, - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + { persistentVolumeName: persistentVolumeName }, - mixin:: {}, - }, - // VolumeAttachmentSpec is the specification of a VolumeAttachment request. - volumeAttachmentSpec:: { - new():: {}, - // Attacher indicates the name of the volume driver that MUST handle this request. This is the name returned by GetPluginName(). - withAttacher(attacher):: self + { attacher: attacher }, - // The node that the volume should be attached to. - withNodeName(nodeName):: self + { nodeName: nodeName }, - mixin:: { - // Source represents the volume that should be attached. - source:: { - local __sourceMixin(source) = { source+: source }, - mixinInstance(source):: __sourceMixin(source), - // Name of the persistent volume to attach. - withPersistentVolumeName(persistentVolumeName):: self + __sourceMixin({ persistentVolumeName: persistentVolumeName }), - }, - sourceType:: hidden.storage.v1beta1.volumeAttachmentSource, - }, - }, - // VolumeAttachmentStatus is the status of a VolumeAttachment request. - volumeAttachmentStatus:: { - new():: {}, - // Indicates the volume is successfully attached. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttached(attached):: self + { attached: attached }, - // Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttachmentMetadata(attachmentMetadata):: self + { attachmentMetadata: attachmentMetadata }, - // Upon successful attach, this field is populated with any information returned by the attach operation that must be passed into subsequent WaitForAttach or Mount calls. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - withAttachmentMetadataMixin(attachmentMetadata):: self + { attachmentMetadata+: attachmentMetadata }, - mixin:: { - // The last error encountered during attach operation, if any. This field must only be set by the entity completing the attach operation, i.e. the external-attacher. - attachError:: { - local __attachErrorMixin(attachError) = { attachError+: attachError }, - mixinInstance(attachError):: __attachErrorMixin(attachError), - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + __attachErrorMixin({ message: message }), - // Time the error was encountered. - withTime(time):: self + __attachErrorMixin({ time: time }), - }, - attachErrorType:: hidden.storage.v1beta1.volumeError, - // The last error encountered during detach operation, if any. This field must only be set by the entity completing the detach operation, i.e. the external-attacher. - detachError:: { - local __detachErrorMixin(detachError) = { detachError+: detachError }, - mixinInstance(detachError):: __detachErrorMixin(detachError), - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + __detachErrorMixin({ message: message }), - // Time the error was encountered. - withTime(time):: self + __detachErrorMixin({ time: time }), - }, - detachErrorType:: hidden.storage.v1beta1.volumeError, - }, - }, - // VolumeError captures an error encountered during a volume operation. - volumeError:: { - new():: {}, - // String detailing the error encountered during Attach or Detach operation. This string may be logged, so it should not contain sensitive information. - withMessage(message):: self + { message: message }, - // Time the error was encountered. - withTime(time):: self + { time: time }, - mixin:: {}, - }, - }, - }, - }, -} \ No newline at end of file diff --git a/example/tk/vendor/ksonnet.beta.4 b/example/tk/vendor/ksonnet.beta.4 deleted file mode 120000 index b753867f7cc..00000000000 --- a/example/tk/vendor/ksonnet.beta.4 +++ /dev/null @@ -1 +0,0 @@ -github.com/ksonnet/ksonnet-lib/ksonnet.beta.4 \ No newline at end of file diff --git a/operations/jsonnet/single-binary/tempo.libsonnet b/operations/jsonnet/single-binary/tempo.libsonnet index ebb1c352c6a..293f031e891 100644 --- a/operations/jsonnet/single-binary/tempo.libsonnet +++ b/operations/jsonnet/single-binary/tempo.libsonnet @@ -23,11 +23,14 @@ pvc.mixin.spec.resources .withRequests({ storage: $._config.pvc_size }) + pvc.mixin.spec - .withAccessModes(['ReadWriteOnce']) + .withAccessModes(['ReadWriteOnce']) + + pvc.mixin.spec .withStorageClassName($._config.pvc_storage_class) + pvc.mixin.metadata - .withLabels({ app: 'tempo' }) - .withNamespace($._config.namespace) + .withLabels({ app: 'tempo' }) + + pvc.mixin.metadata + .withNamespace($._config.namespace) + + pvc.mixin.metadata .withName(tempo_data_volume) + { kind: 'PersistentVolumeClaim', apiVersion: 'v1' }, @@ -82,6 +85,10 @@ tempo_service: k.util.serviceFor($.tempo_statefulset) + service.mixin.spec.withPortsMixin([ - servicePort.withName('http').withPort(80).withTargetPort(16686), + servicePort.newNamed( + name='http', + port=80, + targetPort=16686, + ), ]), } From ca05fb550c39ca865e3bc521c0154982db8753a0 Mon Sep 17 00:00:00 2001 From: Mario Rodriguez Date: Mon, 4 Oct 2021 11:00:45 +0200 Subject: [PATCH 2/3] Add versions to memcached and ksonnet-util imports --- example/tk/jsonnetfile.lock.json | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/example/tk/jsonnetfile.lock.json b/example/tk/jsonnetfile.lock.json index 6d396c292d9..424fdb8f987 100644 --- a/example/tk/jsonnetfile.lock.json +++ b/example/tk/jsonnetfile.lock.json @@ -8,7 +8,7 @@ "subdir": "ksonnet-util" } }, - "version": "9631fe778e87213bed195fcdc8d79da4eecb7431", + "version": "37319e21e546579eab2c3ed459e956a262c28f8c", "sum": "OxgtIWL4hjvG0xkMwUzZ7Yjs52zUhLhaVQpwHCbqf8A=" }, { @@ -18,7 +18,7 @@ "subdir": "memcached" } }, - "version": "9631fe778e87213bed195fcdc8d79da4eecb7431", + "version": "37319e21e546579eab2c3ed459e956a262c28f8c", "sum": "dTOeEux3t9bYSqP2L/uCuLo/wUDpCKH4w+4OD9fePUk=" }, { From 20cdeea414446a0f97b9bba9cbbd2048d1dcc673 Mon Sep 17 00:00:00 2001 From: Mario Date: Mon, 4 Oct 2021 11:08:48 +0200 Subject: [PATCH 3/3] Update CHANGELOG.md Co-authored-by: Koenraad Verheyden --- CHANGELOG.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5095fac62ba..86e18b21907 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -51,7 +51,7 @@ * [ENHANCEMENT] Support global denylist and per-tenant allowlist of tags for search data. [#960](https://github.com/grafana/tempo/pull/960) (@annanay25) * [ENHANCEMENT] Add `search_query_timeout` to Querier config. [#984](https://github.com/grafana/tempo/pull/984) (@kvrhdn) * [ENHANCEMENT] Jsonnet: add `$._config.memcached.memory_limit_mb` [#987](https://github.com/grafana/tempo/pull/987) (@kvrhdn) -* [ENHANCEMENT] Upgrade jsonnet-libs to 1.19 and update tk examples [#](https://github.com/grafana/tempo/pull/) (@mapno) +* [ENHANCEMENT] Upgrade jsonnet-libs to 1.19 and update tk examples [#1001](https://github.com/grafana/tempo/pull/1001) (@mapno) * [BUGFIX] Update port spec for GCS docker-compose example [#869](https://github.com/grafana/tempo/pull/869) (@zalegrala) * [BUGFIX] Fix "magic number" errors and other block mishandling when an ingester forcefully shuts down [#937](https://github.com/grafana/tempo/issues/937) (@mdisibio) * [BUGFIX] Fix compactor memory leak [#806](https://github.com/grafana/tempo/pull/806) (@mdisibio)